From d3dbaae89b12eae1162c7d254efbb6a434730a7a Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Tue, 18 Feb 2025 20:48:13 +0000 Subject: [PATCH 1/2] Fix for R50 TF run when mobilenet datasets are present in cache --- script/app-mlperf-inference-mlcommons-python/meta.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/script/app-mlperf-inference-mlcommons-python/meta.yaml b/script/app-mlperf-inference-mlcommons-python/meta.yaml index 0eadfa0ff..7ad40f9bc 100644 --- a/script/app-mlperf-inference-mlcommons-python/meta.yaml +++ b/script/app-mlperf-inference-mlcommons-python/meta.yaml @@ -519,7 +519,7 @@ deps: # Install datasets ## ImageNet (small for tests) - - tags: get,dataset,image-classification,imagenet,preprocessed + - tags: get,dataset,image-classification,imagenet,preprocessed,-_for.mobilenet names: - imagenet-preprocessed enable_if_env: From 348cc498b3653a16a0fb27ef9c10f513ee493487 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Tue, 18 Feb 2025 20:57:23 +0000 Subject: [PATCH 2/2] Cleanup run-resnet50 --- script/run-all-mlperf-models/run-resnet50.sh | 30 +++++--------------- 1 file changed, 7 insertions(+), 23 deletions(-) diff --git a/script/run-all-mlperf-models/run-resnet50.sh b/script/run-all-mlperf-models/run-resnet50.sh index bd810a277..a2144bf0a 100644 --- a/script/run-all-mlperf-models/run-resnet50.sh +++ b/script/run-all-mlperf-models/run-resnet50.sh @@ -35,6 +35,7 @@ function run_test() { run "$5" } power=' --power=yes --adr.mlperf-power-client.power_server=192.168.0.15 --adr.mlperf-power-client.port=4950 ' +power='' #Add your run commands here... find_performance_cmd='mlcr generate-run-cmds,inference,_find-performance \ @@ -51,37 +52,20 @@ submission_cmd_scenario='mlcr generate-run-cmds,inference,_submission --scenari --category=$category --division=$division --quiet --results_dir=$HOME/results_dir \ --skip_submission_generation=yes --execution_mode=valid $power' -readme_cmd_single='mlcr generate-run-cmds,inference,_populate-readme \ ---model=$model --implementation=$implementation --device=$device --backend=$backend \ ---category=$category --division=$division --quiet --results_dir=$HOME/results_dir \ ---skip_submission_generation=yes --execution_mode=valid $power' - -readme_cmd='mlcr generate-run-cmds,inference,_populate-readme,_all-scenarios \ ---model=$model --implementation=$implementation --device=$device --backend=$backend \ ---category=$category --division=$division --quiet --results_dir=$HOME/results_dir \ ---skip_submission_generation=yes --execution_mode=valid $power' # run "$MLC_RUN_CMD" run_test "onnxruntime" "200" "reference" "cpu" "$find_performance_cmd" run_test "tf" "200" "reference" "cpu" "$find_performance_cmd" -run_test "onnxruntime" "10000" "reference" "cuda" "$find_performance_cmd" -run_test "tf" "20000" "reference" "cuda" "$find_performance_cmd" run_test "onnxruntime" "100" "reference" "cpu" "$submission_cmd" run_test "tf" "100" "reference" "cpu" "$submission_cmd" scenario="SingleStream" run_test "tflite" "100" "tflite-cpp" "cpu" "$submission_cmd_scenario --adr.compiler.tags=gcc" run_test "tflite" "100" "tflite-cpp" "cpu" "$submission_cmd_scenario --adr.compiler.tags=gcc --adr.mlperf-inference-implementation.compressed_dataset=on" + + +run_test "onnxruntime" "10000" "reference" "cuda" "$find_performance_cmd" +run_test "tf" "20000" "reference" "cuda" "$find_performance_cmd" run_test "onnxruntime" "100" "reference" "cuda" "$submission_cmd " -scenario="Offline" -run_test "tf" "100" "reference" "cuda" "$submission_cmd_scenario" -scenario="SingleStream" -run_test "tf" "100" "reference" "cuda" "$submission_cmd_scenario" - -run_test "onnxruntime" "100" "reference" "cpu" "$readme_cmd" -run_test "tf" "100" "reference" "cpu" "$readme_cmd" -run_test "tflite" "100" "tflite-cpp" "cpu" "$readme_cmd_single --adr.compiler.tags=gcc --scenario=SingleStream" -run_test "tflite" "100" "tflite-cpp" "cpu" "$readme_cmd_single --adr.compiler.tags=gcc --scenario=SingleStream --adr.mlperf-inference-implementation.compressed_dataset=on" -run_test "onnxruntime" "100" "reference" "cuda" "$readme_cmd --scenario=SingleStream" -run_test "tf" "100" "reference" "cuda" "$readme_cmd_single --scenario=SingleStream" -run_test "tf" "100" "reference" "cuda" "$readme_cmd_single --scenario=Offline" +run_test "tf" "100" "reference" "cuda" "$submission_cmd" +