From ba8b520675b1ca2b30a76a4591ff3bbaff8d1ce2 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Thu, 2 Jan 2025 15:20:56 +0000 Subject: [PATCH 1/2] Remove ulimit memlock option for nvidia-mlperf-inference --- script/app-mlperf-inference/_cm.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/script/app-mlperf-inference/_cm.yaml b/script/app-mlperf-inference/_cm.yaml index 6c939a225..94f0161a9 100644 --- a/script/app-mlperf-inference/_cm.yaml +++ b/script/app-mlperf-inference/_cm.yaml @@ -409,7 +409,7 @@ variations: nvidia-original: docker: interactive: True - extra_run_args: ' --ulimit memlock=-1 --cap-add SYS_ADMIN --cap-add SYS_TIME --security-opt apparmor=unconfined --security-opt seccomp=unconfined' + extra_run_args: ' --cap-add SYS_ADMIN --cap-add SYS_TIME --security-opt apparmor=unconfined --security-opt seccomp=unconfined' base_image: nvcr.io/nvidia/mlperf/mlperf-inference:mlpinf-v3.1-cuda12.2-cudnn8.9-x86_64-ubuntu20.04-l4-public os: "ubuntu" os_version: "20.04" @@ -434,7 +434,7 @@ variations: CM_HOST_OS_VERSION: - 20.04 docker: - extra_run_args: ' --runtime=nvidia --ulimit memlock=-1 --cap-add SYS_ADMIN --cap-add SYS_TIME --security-opt apparmor=unconfined --security-opt seccomp=unconfined' + extra_run_args: ' --runtime=nvidia --cap-add SYS_ADMIN --cap-add SYS_TIME --security-opt apparmor=unconfined --security-opt seccomp=unconfined' default_variations: backend: tensorrt From 8da5fd39dd7c9df108dbf8795936ea21a61437c1 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Thu, 2 Jan 2025 16:01:36 +0000 Subject: [PATCH 2/2] Added an option to override SUT name during submission generation, turn on checker by default --- script/generate-mlperf-inference-submission/_cm.yaml | 4 ++-- script/generate-mlperf-inference-submission/customize.py | 4 ++-- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/script/generate-mlperf-inference-submission/_cm.yaml b/script/generate-mlperf-inference-submission/_cm.yaml index 5e43dd787..8dad87f25 100644 --- a/script/generate-mlperf-inference-submission/_cm.yaml +++ b/script/generate-mlperf-inference-submission/_cm.yaml @@ -101,9 +101,9 @@ post_deps: - 'yes' - true tags: preprocess,mlperf,submission -- enable_if_env: +- skip_if_env: CM_RUN_SUBMISSION_CHECKER: - - 'yes' + - 'no' names: - mlperf-inference-submission-checker - submission-checker diff --git a/script/generate-mlperf-inference-submission/customize.py b/script/generate-mlperf-inference-submission/customize.py index 73cd70052..d3ab67a2b 100644 --- a/script/generate-mlperf-inference-submission/customize.py +++ b/script/generate-mlperf-inference-submission/customize.py @@ -190,7 +190,7 @@ def generate_submission(env, state, inp, submission_division): result_path, 'system_meta.json') # checks for json file containing system meta sut_info = { - "hardware_name": None, + "system_name": None, "implementation": None, "device": None, "framework": None, @@ -276,7 +276,7 @@ def generate_submission(env, state, inp, submission_division): {model: returned_model_name}) if check_dict_filled(sut_info.keys(), sut_info): - system = sut_info["hardware_name"] + system = env.get('CM_HW_NAME', sut_info["system_name"]) implementation = sut_info["implementation"] device = sut_info["device"] framework = sut_info["framework"].replace(" ", "_")