@@ -1000,16 +1000,24 @@ def _generate_prod(dataset, cfg, prdname, prdfunc, dsname, voltime, runinfo=None
10001000 filenames = prdfunc (dataset , prdcfg )
10011001 if isinstance (filenames , str ): # convert to list if needed
10021002 filenames = [filenames ]
1003- if "s3copypath" in prdcfg and filenames is not None : # copy to S3
1004- s3AccessPolicy = (
1005- prdcfg ["s3AccessPolicy" ] if "s3AccessPolicy" in prdcfg else None
1006- )
1003+ if (
1004+ "s3BucketWrite" in prdcfg
1005+ and "s3EndpointWrite" in prdcfg
1006+ and filenames is not None
1007+ ): # copy to S3
1008+ s3AccessPolicy = prdcfg .get ("s3AccessPolicy" , None )
1009+ s3path = prdcfg .get ("s3PathWrite" , None )
10071010 for fname in filenames :
10081011 if (
10091012 prdcfg ["basepath" ] in fname
10101013 ): # only products saved to standard basepath
10111014 write_to_s3 (
1012- fname , prdcfg ["basepath" ], prdcfg ["s3copypath" ], s3AccessPolicy
1015+ fname ,
1016+ prdcfg ["basepath" ],
1017+ prdcfg ["s3EndpointWrite" ],
1018+ prdcfg ["s3BucketWrite" ],
1019+ s3path ,
1020+ s3AccessPolicy ,
10131021 )
10141022 return False
10151023 except Exception as inst :
@@ -1221,6 +1229,9 @@ def _create_cfg_dict(cfgfile):
12211229 if "MFScale" not in cfg :
12221230 cfg .update ({"MFScale" : 0 })
12231231
1232+ if not cfg ["datapath" ]: # empty datapath in case of s3 reading
1233+ cfg ["datapath" ] = ["" for rad in range (cfg ["NumRadars" ])]
1234+
12241235 # parameters necessary to read correctly MF grid binary files
12251236 if "BinFileParams" not in cfg :
12261237 bin_file_params = {
@@ -1333,35 +1344,35 @@ def _create_datacfg_dict(cfg):
13331344 datacfg .update ({"DataTypeIDInFiles" : cfg ["DataTypeIDInFiles" ]})
13341345
13351346 # s3 buckets
1336- if "bucket " in cfg :
1347+ if "s3BucketRead " in cfg :
13371348 try :
1338- datacfg ["s3_key " ] = os .environ ["S3_IN_KEY " ]
1339- datacfg ["s3_secret_key " ] = os .environ ["S3_IN_SECRET " ]
1349+ datacfg ["s3KeyRead " ] = os .environ ["S3_KEY_READ " ]
1350+ datacfg ["s3SecretRead " ] = os .environ ["S3_SECRET_READ " ]
13401351 except KeyError :
13411352 warn (
1342- "Define environment variables S3_IN_KEY and S3_IN_SECRET "
1353+ "Define environment variables S3_KEY_READ and S3_SECRET_READ "
13431354 " to get input data from S3 buckets."
13441355 )
13451356
1346- if "s3path " in cfg :
1347- datacfg .update ({"s3path " : cfg ["s3path " ]})
1357+ if "s3PathRead " in cfg :
1358+ datacfg .update ({"s3PathRead " : cfg ["s3PathRead " ]})
13481359 else :
1349- warn ("Unable to read data from s3 bucket. Define s3path " )
1350- if "s3_url " in cfg :
1351- datacfg .update ({"s3_url " : cfg ["s3_url " ]})
1360+ warn ("Unable to read data from s3 bucket. Define s3PathRead " )
1361+ if "s3EndpointRead " in cfg :
1362+ datacfg .update ({"s3EndpointRead " : cfg ["s3EndpointRead " ]})
13521363 else :
1353- warn ("Unable to read data from s3 bucket. Define s3_url " )
1364+ warn ("Unable to read data from s3 bucket. Define s3EndpointRead " )
13541365
13551366 if "rm_s3_file" in cfg :
13561367 datacfg .update ({"rm_s3_file" : cfg ["rm_s3_file" ]})
13571368
13581369 if (
1359- "s3path " in datacfg
1360- and "s3_url " in datacfg
1361- and "s3_key " in datacfg
1362- and "s3_secret_key " in datacfg
1370+ "s3PathRead " in datacfg
1371+ and "s3EndpointRead " in datacfg
1372+ and "s3KeyRead " in datacfg
1373+ and "s3SecretRead " in datacfg
13631374 ):
1364- datacfg .update ({"bucket " : cfg ["bucket " ]})
1375+ datacfg .update ({"s3BucketRead " : cfg ["s3BucketRead " ]})
13651376
13661377 # Modify size of radar or radar spectra object
13671378 datacfg .update ({"elmin" : cfg .get ("elmin" , None )})
@@ -1582,11 +1593,14 @@ def _create_prdcfg_dict(cfg, dataset, product, voltime, runinfo=None):
15821593 prdcfg .update ({"imgformat" : cfg ["imgformat" ]})
15831594 prdcfg .update ({"RadarName" : cfg ["RadarName" ]})
15841595
1585- if "s3copypath" in cfg :
1586- prdcfg .update ({"s3copypath" : cfg ["s3copypath" ]})
1596+ if "s3EndpointWrite" in cfg :
1597+ prdcfg .update ({"s3EndpointWrite" : cfg ["s3EndpointWrite" ]})
1598+ if "s3BucketWrite" in cfg :
1599+ prdcfg .update ({"s3BucketWrite" : cfg ["s3BucketWrite" ]})
1600+ if "s3PathWrite" in cfg :
1601+ prdcfg .update ({"s3PathWrite" : cfg ["s3PathWrite" ]})
15871602 if "s3AccessPolicy" in cfg :
15881603 prdcfg .update ({"s3AccessPolicy" : cfg ["s3AccessPolicy" ]})
1589-
15901604 if "RadarBeamwidth" in cfg :
15911605 prdcfg .update ({"RadarBeamwidth" : cfg ["RadarBeamwidth" ]})
15921606 if "ppiImageConfig" in cfg :
@@ -1641,9 +1655,13 @@ def _get_datatype_list(cfg, radarnr="RADAR001"):
16411655 if "datatype" not in cfg [dataset ]:
16421656 continue
16431657 if isinstance (cfg [dataset ]["datatype" ], str ):
1644- (radarnr_descr , datagroup , datatype_aux , dataset_save , product_save ) = (
1645- get_datatype_fields (cfg [dataset ]["datatype" ])
1646- )
1658+ (
1659+ radarnr_descr ,
1660+ datagroup ,
1661+ datatype_aux ,
1662+ dataset_save ,
1663+ product_save ,
1664+ ) = get_datatype_fields (cfg [dataset ]["datatype" ])
16471665 if datagroup != "PROC" and radarnr_descr == radarnr :
16481666 if (dataset_save is None ) and (product_save is None ):
16491667 datatypesdescr .add (
@@ -1667,9 +1685,13 @@ def _get_datatype_list(cfg, radarnr="RADAR001"):
16671685 )
16681686 else :
16691687 for datatype in cfg [dataset ]["datatype" ]:
1670- (radarnr_descr , datagroup , datatype_aux , dataset_save , product_save ) = (
1671- get_datatype_fields (datatype )
1672- )
1688+ (
1689+ radarnr_descr ,
1690+ datagroup ,
1691+ datatype_aux ,
1692+ dataset_save ,
1693+ product_save ,
1694+ ) = get_datatype_fields (datatype )
16731695 if datagroup != "PROC" and radarnr_descr == radarnr :
16741696 if dataset_save is None and product_save is None :
16751697 datatypesdescr .add (
@@ -1799,7 +1821,7 @@ def _get_masterfile_list(datatypesdescr, starttimes, endtimes, datacfg, scan_lis
17991821 )
18001822 return [], None , None
18011823
1802- if "bucket " in datacfg :
1824+ if "s3BucketRead " in datacfg :
18031825 masterfilelist = get_file_list_s3 (
18041826 masterdatatypedescr , starttimes , endtimes , datacfg , scan = masterscan
18051827 )
0 commit comments