diff --git a/script/get-lib-armnn/meta.yaml b/script/get-lib-armnn/meta.yaml index 6c33e97ed..ff71bf95a 100644 --- a/script/get-lib-armnn/meta.yaml +++ b/script/get-lib-armnn/meta.yaml @@ -3,7 +3,7 @@ automation_alias: script automation_uid: 5b4e0237da074764 cache: true category: Detection or installation of tools and artifacts -default_version: '23.11' +default_version: '25.02' deps: - tags: detect,os env: diff --git a/script/run-all-mlperf-models/run-mobilenet-models.sh b/script/run-all-mlperf-models/run-mobilenet-models.sh index 9ad2bc7cd..ad24ed87d 100644 --- a/script/run-all-mlperf-models/run-mobilenet-models.sh +++ b/script/run-all-mlperf-models/run-mobilenet-models.sh @@ -24,7 +24,8 @@ function run() { } POWER=" --power=yes --adr.mlperf-power-client.power_server=192.168.0.15 --adr.mlperf-power-client.port=4940 " POWER="" -extra_option="" +#extra_option=" --minimize_disk_usage=yes" +extra_option=" --minimize_disk_usage=no" extra_tags="" #extra_option=" --adr.mlperf-inference-implementation.compressed_dataset=on" #extra_tags=",_only-fp32" @@ -41,7 +42,8 @@ run "mlcr run,mobilenet-models,_tflite,_armnn,_neon$extra_tags \ ${extra_option} " -extra_option=" --adr.mlperf-inference-implementation.compressed_dataset=on" +extra_option=" $extra_option --adr.mlperf-inference-implementation.compressed_dataset=on" +extra_tag=",_only-fp32" run "mlcr run,mobilenet-models,_tflite$extra_tags \ --adr.compiler.tags=gcc \ ${extra_option} " diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 9601af651..a7e3d5ee0 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -195,7 +195,7 @@ def preprocess(i): return r importlib.reload(mlc.action) - if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): + if is_true(env.get('MLC_MINIMIZE_DISK_USAGE', '')): r = cache_action.access(clean_input) if r['return'] > 0: print(r) @@ -206,12 +206,14 @@ def preprocess(i): if is_true(env.get('MLC_TEST_ONE_RUN', '')): return {'return': 0} + ''' r = cache_action.access(clean_input) if r['return'] > 0: print(r) # return r else: importlib.reload(mlc.action) + ''' return {'return': 0} diff --git a/script/run-mlperf-inference-mobilenet-models/meta.yaml b/script/run-mlperf-inference-mobilenet-models/meta.yaml index a6a42760a..813b1ef93 100644 --- a/script/run-mlperf-inference-mobilenet-models/meta.yaml +++ b/script/run-mlperf-inference-mobilenet-models/meta.yaml @@ -9,7 +9,7 @@ default_env: MLC_MLPERF_RUN_INT8: 'yes' MLC_MLPERF_RUN_MOBILENETS: 'no' MLC_USE_DATASET_FROM_HOST: 'yes' - MLC_MINIMIZE_DISK_SPACE: 'yes' + MLC_MINIMIZE_DISK_USAGE: 'no' deps: - tags: get,sys-utils-mlc docker: @@ -35,6 +35,7 @@ docker: imagenet_path: MLC_DATASET_IMAGENET_PATH results_dir: MLC_MLPERF_INFERENCE_RESULTS_DIR submission_dir: MLC_MLPERF_INFERENCE_SUBMISSION_DIR + minimize_disk_usage: MLC_MINIMIZE_DISK_USAGE fake_run_deps: false pre_run_cmds: - mlc pull repo