From 0af474cfba5b494ef6ee111fb362c3b46e5f0115 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Mon, 24 Feb 2025 17:09:49 +0000 Subject: [PATCH 01/16] Fix boolean usage in run-mobilenet-models --- .../customize.py | 17 +++++++++-------- 1 file changed, 9 insertions(+), 8 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 2588aca8e..53602cac2 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -107,17 +107,17 @@ def preprocess(i): execution_mode = "test" precisions = [] - if env.get('MLC_MLPERF_RUN_FP32', '') == "yes": + if is_true(env.get('MLC_MLPERF_RUN_FP32', '') == "yes": precisions.append("fp32") - if env.get('MLC_MLPERF_RUN_INT8', '') == "yes": + if is_true(env.get('MLC_MLPERF_RUN_INT8', '') == "yes": precisions.append("uint8") implementation_tags = [] - if env.get('MLC_MLPERF_USE_ARMNN_LIBRARY', '') == "yes": + if is_true(env.get('MLC_MLPERF_USE_ARMNN_LIBRARY', '')): implementation_tags.append("_armnn") - if env.get('MLC_MLPERF_TFLITE_ARMNN_NEON', '') == "yes": + if is_true(env.get('MLC_MLPERF_TFLITE_ARMNN_NEON', '')): implementation_tags.append("_use-neon") - if env.get('MLC_MLPERF_TFLITE_ARMNN_OPENCL', '') == "yes": + if is_true(env.get('MLC_MLPERF_TFLITE_ARMNN_OPENCL', '')): implementation_tags.append("_use-opencl") implementation_tags_string = ",".join(implementation_tags) @@ -200,15 +200,16 @@ def preprocess(i): if r['return'] > 0: return r - if env.get('MLC_TEST_ONE_RUN', '') == "yes": - return {'return': 0} - if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): r = mlc.access(clean_input) if r['return'] > 0: print(r) # return r + if is_true(env.get('MLC_TEST_ONE_RUN', '')): + return {'return': 0} + + r = mlc.access(clean_input) if r['return'] > 0: print(r) From 9e031d509ede69c7181a958bb4971c32c9758e4f Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 24 Feb 2025 17:10:07 +0000 Subject: [PATCH 02/16] [Automated Commit] Format Codebase [skip ci] --- .../customize.py | 34 +++++++++---------- 1 file changed, 17 insertions(+), 17 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 53602cac2..7888094f9 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -112,17 +112,17 @@ def preprocess(i): if is_true(env.get('MLC_MLPERF_RUN_INT8', '') == "yes": precisions.append("uint8") - implementation_tags = [] + implementation_tags=[] if is_true(env.get('MLC_MLPERF_USE_ARMNN_LIBRARY', '')): implementation_tags.append("_armnn") if is_true(env.get('MLC_MLPERF_TFLITE_ARMNN_NEON', '')): implementation_tags.append("_use-neon") if is_true(env.get('MLC_MLPERF_TFLITE_ARMNN_OPENCL', '')): implementation_tags.append("_use-opencl") - implementation_tags_string = ",".join(implementation_tags) + implementation_tags_string=",".join(implementation_tags) - inp = i['input'] - clean_input = { + inp=i['input'] + clean_input={ 'action': 'rm', 'target': 'cache', 'tags': 'get,preprocessed,dataset,_for.mobilenet', @@ -139,9 +139,9 @@ def preprocess(i): continue if model == "efficientnet" and precision == "uint8": - precision = "int8" + precision="int8" - mlc_input = { + mlc_input={ 'action': 'run', 'target': 'script', 'tags': f'run-mlperf,mlperf,inference,{var}', @@ -166,42 +166,42 @@ def preprocess(i): } if add_deps_recursive: # script automation will merge adr and add_deps_recursive - mlc_input['add_deps_recursive'] = add_deps_recursive + mlc_input['add_deps_recursive']=add_deps_recursive if adr: utils.merge_dicts( {'dict1': mlc_input['adr'], 'dict2': adr, 'append_lists': True, 'append_unique': True}) if env.get('MLC_MLPERF_INFERENCE_RESULTS_DIR', '') != '': - mlc_input['results_dir'] = env['MLC_MLPERF_INFERENCE_RESULTS_DIR'] + mlc_input['results_dir']=env['MLC_MLPERF_INFERENCE_RESULTS_DIR'] if env.get('MLC_MLPERF_INFERENCE_SUBMISSION_DIR', '') != '': - mlc_input['submission_dir'] = env['MLC_MLPERF_INFERENCE_SUBMISSION_DIR'] + mlc_input['submission_dir']=env['MLC_MLPERF_INFERENCE_SUBMISSION_DIR'] if is_true(env.get('MLC_MLPERF_FIND_PERFORMANCE_MODE', '')) and not is_true(env.get( 'MLC_MLPERF_NO_RERUN', '')): - mlc_input['rerun'] = True + mlc_input['rerun']=True if is_true(env.get('MLC_MLPERF_POWER', '')): - mlc_input['power'] = 'yes' + mlc_input['power']='yes' if is_true(env.get('MLC_MLPERF_ACCURACY_MODE', '')): - mlc_input['mode'] = 'accuracy' + mlc_input['mode']='accuracy' print(mlc_input) - r = mlc.access(mlc_input) + r=mlc.access(mlc_input) if r['return'] > 0: return r if is_true(env.get('MLC_MLPERF_PERFORMANCE_MODE', '')): - mlc_input['mode'] = 'performance' + mlc_input['mode']='performance' print(mlc_input) - r = mlc.access(mlc_input) + r=mlc.access(mlc_input) if r['return'] > 0: return r if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): - r = mlc.access(clean_input) + r=mlc.access(clean_input) if r['return'] > 0: print(r) # return r @@ -210,7 +210,7 @@ def preprocess(i): return {'return': 0} - r = mlc.access(clean_input) + r=mlc.access(clean_input) if r['return'] > 0: print(r) # return r From c4cf1f508ae1e5cf1433d7577c1cde4f3cc7ad8e Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Mon, 24 Feb 2025 17:11:33 +0000 Subject: [PATCH 03/16] Fix boolean usage in run-mobilenet-models --- script/run-mlperf-inference-mobilenet-models/customize.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 7888094f9..9f73f01b6 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -107,9 +107,9 @@ def preprocess(i): execution_mode = "test" precisions = [] - if is_true(env.get('MLC_MLPERF_RUN_FP32', '') == "yes": + if is_true(env.get('MLC_MLPERF_RUN_FP32', '')): precisions.append("fp32") - if is_true(env.get('MLC_MLPERF_RUN_INT8', '') == "yes": + if is_true(env.get('MLC_MLPERF_RUN_INT8', '')): precisions.append("uint8") implementation_tags=[] From 5b84eea77b21113302882ee582d097b6629937c0 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 24 Feb 2025 17:11:55 +0000 Subject: [PATCH 04/16] [Automated Commit] Format Codebase [skip ci] --- .../customize.py | 35 +++++++++---------- 1 file changed, 17 insertions(+), 18 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 9f73f01b6..396351d89 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -112,17 +112,17 @@ def preprocess(i): if is_true(env.get('MLC_MLPERF_RUN_INT8', '')): precisions.append("uint8") - implementation_tags=[] + implementation_tags = [] if is_true(env.get('MLC_MLPERF_USE_ARMNN_LIBRARY', '')): implementation_tags.append("_armnn") if is_true(env.get('MLC_MLPERF_TFLITE_ARMNN_NEON', '')): implementation_tags.append("_use-neon") if is_true(env.get('MLC_MLPERF_TFLITE_ARMNN_OPENCL', '')): implementation_tags.append("_use-opencl") - implementation_tags_string=",".join(implementation_tags) + implementation_tags_string = ",".join(implementation_tags) - inp=i['input'] - clean_input={ + inp = i['input'] + clean_input = { 'action': 'rm', 'target': 'cache', 'tags': 'get,preprocessed,dataset,_for.mobilenet', @@ -139,9 +139,9 @@ def preprocess(i): continue if model == "efficientnet" and precision == "uint8": - precision="int8" + precision = "int8" - mlc_input={ + mlc_input = { 'action': 'run', 'target': 'script', 'tags': f'run-mlperf,mlperf,inference,{var}', @@ -166,42 +166,42 @@ def preprocess(i): } if add_deps_recursive: # script automation will merge adr and add_deps_recursive - mlc_input['add_deps_recursive']=add_deps_recursive + mlc_input['add_deps_recursive'] = add_deps_recursive if adr: utils.merge_dicts( {'dict1': mlc_input['adr'], 'dict2': adr, 'append_lists': True, 'append_unique': True}) if env.get('MLC_MLPERF_INFERENCE_RESULTS_DIR', '') != '': - mlc_input['results_dir']=env['MLC_MLPERF_INFERENCE_RESULTS_DIR'] + mlc_input['results_dir'] = env['MLC_MLPERF_INFERENCE_RESULTS_DIR'] if env.get('MLC_MLPERF_INFERENCE_SUBMISSION_DIR', '') != '': - mlc_input['submission_dir']=env['MLC_MLPERF_INFERENCE_SUBMISSION_DIR'] + mlc_input['submission_dir'] = env['MLC_MLPERF_INFERENCE_SUBMISSION_DIR'] if is_true(env.get('MLC_MLPERF_FIND_PERFORMANCE_MODE', '')) and not is_true(env.get( 'MLC_MLPERF_NO_RERUN', '')): - mlc_input['rerun']=True + mlc_input['rerun'] = True if is_true(env.get('MLC_MLPERF_POWER', '')): - mlc_input['power']='yes' + mlc_input['power'] = 'yes' if is_true(env.get('MLC_MLPERF_ACCURACY_MODE', '')): - mlc_input['mode']='accuracy' + mlc_input['mode'] = 'accuracy' print(mlc_input) - r=mlc.access(mlc_input) + r = mlc.access(mlc_input) if r['return'] > 0: return r if is_true(env.get('MLC_MLPERF_PERFORMANCE_MODE', '')): - mlc_input['mode']='performance' + mlc_input['mode'] = 'performance' print(mlc_input) - r=mlc.access(mlc_input) + r = mlc.access(mlc_input) if r['return'] > 0: return r if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): - r=mlc.access(clean_input) + r = mlc.access(clean_input) if r['return'] > 0: print(r) # return r @@ -209,8 +209,7 @@ def preprocess(i): if is_true(env.get('MLC_TEST_ONE_RUN', '')): return {'return': 0} - - r=mlc.access(clean_input) + r = mlc.access(clean_input) if r['return'] > 0: print(r) # return r From a7dabd271c246a141a387dac3340c0d632d38a02 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Mon, 24 Feb 2025 18:35:49 +0000 Subject: [PATCH 05/16] Update test-scc24-sdxl.yaml --- .github/workflows/test-scc24-sdxl.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/test-scc24-sdxl.yaml b/.github/workflows/test-scc24-sdxl.yaml index ffe814ba4..77e3b044e 100644 --- a/.github/workflows/test-scc24-sdxl.yaml +++ b/.github/workflows/test-scc24-sdxl.yaml @@ -2,7 +2,7 @@ name: MLPerf inference SDXL (SCC) on: schedule: - - cron: "34 19 * * *" + - cron: "34 19 1 * *" jobs: build_reference: From 4a191c3d0a9fd24536af3f4b105ac68e335b3c72 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Mon, 24 Feb 2025 20:10:56 +0000 Subject: [PATCH 06/16] Workaround for cache not working in mobilenet run --- .../run-mlperf-inference-mobilenet-models/customize.py | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 396351d89..e2f79643c 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -2,7 +2,8 @@ import os import sys from utils import * - +import mlc +import importlib def preprocess(i): @@ -17,7 +18,7 @@ def preprocess(i): adr = i['input'].get('adr') automation = i['automation'] - mlc = i['automation'].action_object + #mlc = i['automation'].action_object quiet = (env.get('MLC_QUIET', False) == 'yes') verbose = (env.get('MLC_VERBOSE', False) == 'yes') @@ -192,6 +193,8 @@ def preprocess(i): if r['return'] > 0: return r + importlib.reload(mlc) + if is_true(env.get('MLC_MLPERF_PERFORMANCE_MODE', '')): mlc_input['mode'] = 'performance' @@ -200,10 +203,13 @@ def preprocess(i): if r['return'] > 0: return r + importlib.reload(mlc) + if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): r = mlc.access(clean_input) if r['return'] > 0: print(r) + importlib.reload(mlc) # return r if is_true(env.get('MLC_TEST_ONE_RUN', '')): From 0724a14f37e227a8448543f34269f98d9e78746d Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 24 Feb 2025 20:11:18 +0000 Subject: [PATCH 07/16] [Automated Commit] Format Codebase [skip ci] --- script/run-mlperf-inference-mobilenet-models/customize.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index e2f79643c..f938f3ee0 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -5,6 +5,7 @@ import mlc import importlib + def preprocess(i): os_info = i['os_info'] @@ -18,7 +19,7 @@ def preprocess(i): adr = i['input'].get('adr') automation = i['automation'] - #mlc = i['automation'].action_object + # mlc = i['automation'].action_object quiet = (env.get('MLC_QUIET', False) == 'yes') verbose = (env.get('MLC_VERBOSE', False) == 'yes') From 59d5918708bc94e8aae5aac1ef77249cd298548e Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Mon, 24 Feb 2025 21:12:36 +0000 Subject: [PATCH 08/16] Revert Workaround for cache not working in mobilenet run --- script/run-mlperf-inference-mobilenet-models/customize.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index f938f3ee0..17ddef3d0 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -3,7 +3,6 @@ import sys from utils import * import mlc -import importlib def preprocess(i): @@ -194,7 +193,6 @@ def preprocess(i): if r['return'] > 0: return r - importlib.reload(mlc) if is_true(env.get('MLC_MLPERF_PERFORMANCE_MODE', '')): mlc_input['mode'] = 'performance' @@ -204,13 +202,11 @@ def preprocess(i): if r['return'] > 0: return r - importlib.reload(mlc) if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): r = mlc.access(clean_input) if r['return'] > 0: print(r) - importlib.reload(mlc) # return r if is_true(env.get('MLC_TEST_ONE_RUN', '')): From 6856e26e479839a793ad3546a49b7942a5ee17bf Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 24 Feb 2025 21:12:53 +0000 Subject: [PATCH 09/16] [Automated Commit] Format Codebase [skip ci] --- script/run-mlperf-inference-mobilenet-models/customize.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 17ddef3d0..8bc2905ca 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -193,7 +193,6 @@ def preprocess(i): if r['return'] > 0: return r - if is_true(env.get('MLC_MLPERF_PERFORMANCE_MODE', '')): mlc_input['mode'] = 'performance' @@ -202,7 +201,6 @@ def preprocess(i): if r['return'] > 0: return r - if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): r = mlc.access(clean_input) if r['return'] > 0: From eae969d18725a32087ace83214dd91b0997e969f Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Mon, 24 Feb 2025 22:27:43 +0000 Subject: [PATCH 10/16] Revert Workaround for cache not working in mobilenet run --- script/run-mlperf-inference-mobilenet-models/customize.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 8bc2905ca..396351d89 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -2,7 +2,6 @@ import os import sys from utils import * -import mlc def preprocess(i): @@ -18,7 +17,7 @@ def preprocess(i): adr = i['input'].get('adr') automation = i['automation'] - # mlc = i['automation'].action_object + mlc = i['automation'].action_object quiet = (env.get('MLC_QUIET', False) == 'yes') verbose = (env.get('MLC_VERBOSE', False) == 'yes') From 10e53a4b078d739fe5930f70cb21efdf2b47240c Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Mon, 24 Feb 2025 22:36:10 +0000 Subject: [PATCH 11/16] Fix access function for cache in mobilenet run --- script/run-mlperf-inference-mobilenet-models/customize.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 396351d89..b57daf283 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -18,6 +18,7 @@ def preprocess(i): automation = i['automation'] mlc = i['automation'].action_object + cache_action = i['automation'].cache_action quiet = (env.get('MLC_QUIET', False) == 'yes') verbose = (env.get('MLC_VERBOSE', False) == 'yes') @@ -201,7 +202,7 @@ def preprocess(i): return r if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): - r = mlc.access(clean_input) + r = cache_action.access(clean_input) if r['return'] > 0: print(r) # return r @@ -209,7 +210,7 @@ def preprocess(i): if is_true(env.get('MLC_TEST_ONE_RUN', '')): return {'return': 0} - r = mlc.access(clean_input) + r = cache_action.access(clean_input) if r['return'] > 0: print(r) # return r From f5baef13e8074157a6869b599790a3d8b840a701 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Tue, 25 Feb 2025 10:06:32 +0000 Subject: [PATCH 12/16] Fix efficientnet runs --- script/run-mlperf-inference-mobilenet-models/customize.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index b57daf283..99a7524e8 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -62,7 +62,7 @@ def preprocess(i): models['mobilenet']['v3'] = models_all['mobilenet']['v3'] elif is_true(env.get('MLC_MLPERF_RUN_MOBILENETS', '')): models['mobilenet'] = models_all['mobilenet'] - elif is_true(env.get('MLC_MLPERF_RUN_EFFICIENTNETS', '')): + if is_true(env.get('MLC_MLPERF_RUN_EFFICIENTNETS', '')): models['efficientnet'] = models_all['efficientnet'] variation_strings = {} From 5f49da685d9e6f18783f27d30124a5336f566320 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Tue, 25 Feb 2025 10:27:47 +0000 Subject: [PATCH 13/16] Fix double run of perf and accuracy modes for mobilenets --- .../customize.py | 18 ++++-------------- 1 file changed, 4 insertions(+), 14 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 99a7524e8..371b2dc69 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -186,20 +186,10 @@ def preprocess(i): if is_true(env.get('MLC_MLPERF_POWER', '')): mlc_input['power'] = 'yes' - if is_true(env.get('MLC_MLPERF_ACCURACY_MODE', '')): - mlc_input['mode'] = 'accuracy' - print(mlc_input) - r = mlc.access(mlc_input) - if r['return'] > 0: - return r - - if is_true(env.get('MLC_MLPERF_PERFORMANCE_MODE', '')): - mlc_input['mode'] = 'performance' - - print(mlc_input) - r = mlc.access(mlc_input) - if r['return'] > 0: - return r + print(mlc_input) + r = mlc.access(mlc_input) + if r['return'] > 0: + return r if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): r = cache_action.access(clean_input) From c76565ca336df5f9ffffc1d364784913d4961fb0 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Tue, 25 Feb 2025 10:38:09 +0000 Subject: [PATCH 14/16] Workaround for mobilenet cache cleaning --- .../customize.py | 11 ++++++++--- 1 file changed, 8 insertions(+), 3 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index 371b2dc69..a1a4c5d2c 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -2,7 +2,8 @@ import os import sys from utils import * - +import mlc +import importlib def preprocess(i): @@ -17,8 +18,9 @@ def preprocess(i): adr = i['input'].get('adr') automation = i['automation'] - mlc = i['automation'].action_object - cache_action = i['automation'].cache_action + #mlc = i['automation'].action_object + #cache_action = i['automation'].cache_action + cache_action = mlc quiet = (env.get('MLC_QUIET', False) == 'yes') verbose = (env.get('MLC_VERBOSE', False) == 'yes') @@ -190,6 +192,7 @@ def preprocess(i): r = mlc.access(mlc_input) if r['return'] > 0: return r + importlib.reload(mlc.action) if is_true(env.get('MLC_MINIMIZE_DISK_SPACE', '')): r = cache_action.access(clean_input) @@ -204,6 +207,8 @@ def preprocess(i): if r['return'] > 0: print(r) # return r + else: + importlib.reload(mlc.action) return {'return': 0} From e0e8a8e34c4d8e1b7901d1ecfd0a66e275321fae Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Tue, 25 Feb 2025 10:38:25 +0000 Subject: [PATCH 15/16] [Automated Commit] Format Codebase [skip ci] --- script/run-mlperf-inference-mobilenet-models/customize.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/script/run-mlperf-inference-mobilenet-models/customize.py b/script/run-mlperf-inference-mobilenet-models/customize.py index a1a4c5d2c..fb3238ca2 100644 --- a/script/run-mlperf-inference-mobilenet-models/customize.py +++ b/script/run-mlperf-inference-mobilenet-models/customize.py @@ -5,6 +5,7 @@ import mlc import importlib + def preprocess(i): os_info = i['os_info'] @@ -18,8 +19,8 @@ def preprocess(i): adr = i['input'].get('adr') automation = i['automation'] - #mlc = i['automation'].action_object - #cache_action = i['automation'].cache_action + # mlc = i['automation'].action_object + # cache_action = i['automation'].cache_action cache_action = mlc quiet = (env.get('MLC_QUIET', False) == 'yes') From d41c892380bea644e29fbf892e6ce7bb60c7154e Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Tue, 25 Feb 2025 10:40:54 +0000 Subject: [PATCH 16/16] Update test-scc24-sdxl.yaml --- .github/workflows/test-scc24-sdxl.yaml | 1 + 1 file changed, 1 insertion(+) diff --git a/.github/workflows/test-scc24-sdxl.yaml b/.github/workflows/test-scc24-sdxl.yaml index 77e3b044e..3151c8183 100644 --- a/.github/workflows/test-scc24-sdxl.yaml +++ b/.github/workflows/test-scc24-sdxl.yaml @@ -1,5 +1,6 @@ name: MLPerf inference SDXL (SCC) + on: schedule: - cron: "34 19 1 * *"