Skip to content

Commit 0e647d7

Browse files
authored
Merge pull request #203 from mlcommons/dev
Merge Dev
2 parents 2ac0b7b + f798762 commit 0e647d7

File tree

42 files changed

+204
-69
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

42 files changed

+204
-69
lines changed

.github/workflows/test-mlc-script-features.yml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -113,7 +113,9 @@ jobs:
113113
114114
mlcr --tags=install,python-venv --version=3.10.8 --name=mlperf --quiet
115115
116-
mlcr --tags=run,mlperf,inference,_submission,_short --adr.python.name=mlperf --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --submitter=MLCommons --implementation=cpp --hw_name=default --model=retinanet --backend=onnxruntime --device=cpu --scenario=Offline --quiet
116+
export MLC_SCRIPT_EXTRA_CMD="--adr.python.name=mlperf"
117+
118+
mlcr --tags=run,mlperf,inference,_submission,_short --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --submitter=MLCommons --implementation=cpp --hw_name=default --model=retinanet --backend=onnxruntime --device=cpu --scenario=Offline --quiet
117119
118120
# Step for Linux/MacOS
119121
- name: Randomly Execute Step (Linux/MacOS)

.github/workflows/test-mlperf-inference-abtf-poc.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@ on:
1010

1111
jobs:
1212
build:
13+
name: Test MLPerf Inference ABTF POC
1314
runs-on: ${{ matrix.os }}
1415
strategy:
1516
fail-fast: false

.github/workflows/test-mlperf-inference-bert-deepsparse-tf-onnxruntime-pytorch.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@ on:
1010

1111
jobs:
1212
build:
13+
name: MLPerf Inference Bert ${{ matrix.backend }} on ${{ matrix.os }}
1314
runs-on: ${{ matrix.os }}
1415
strategy:
1516
fail-fast: false

.github/workflows/test-mlperf-inference-mlcommons-cpp-resnet50.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@ on:
1010

1111
jobs:
1212
build:
13+
name: MLPerf inference MLCommons C++ ResNet50
1314
runs-on: ${{ matrix.os }}
1415
strategy:
1516
fail-fast: false

.github/workflows/test-mlperf-inference-rgat.yml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,8 @@ on:
99
- '!**.md'
1010

1111
jobs:
12-
build:
12+
rgat-inference-run:
13+
name: ${{ matrix.os }} - ${{ matrix.backend }} - ${{ matrix.implementation }}
1314
runs-on: ${{ matrix.os }}
1415
strategy:
1516
fail-fast: false

.github/workflows/test-mlperf-inference-tvm-resnet50.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@ on:
1010

1111
jobs:
1212
tvm-run:
13+
name: TVM ResNet50
1314
runs-on: ubuntu-latest
1415
strategy:
1516
fail-fast: false

.github/workflows/test-mlperf-loadgen-onnx-huggingface-bert-fp32-squad.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ on:
1010

1111
jobs:
1212
build:
13-
13+
name: Test MLPerf loadgen with HuggingFace bert onnx fp32 squad model on Python ${{ matrix.python-version }}
1414
runs-on: ubuntu-latest
1515
strategy:
1616
fail-fast: false

.github/workflows/test-nvidia-mlperf-inference-implementations.yml

Lines changed: 4 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -17,12 +17,14 @@ jobs:
1717
strategy:
1818
fail-fast: false
1919
matrix:
20-
system: [ "GO-spr", "phoenix"]
20+
system: [ "GO-spr", "phoenix", "GO-i9" ]
2121
# system: [ "mlc-server" ]
2222
python-version: [ "3.12" ]
2323
model: [ "resnet50", "retinanet", "bert-99", "bert-99.9", "gptj-99.9", "3d-unet-99.9", "sdxl" ]
2424
exclude:
2525
- model: gptj-99.9
26+
- system: phoenix
27+
- system: GO-i9
2628

2729
steps:
2830
- name: Test MLPerf Inference NVIDIA ${{ matrix.model }}
@@ -43,12 +45,7 @@ jobs:
4345
gpu_name=rtx_4090
4446
docker_string=" --docker"
4547
fi
46-
if [ "${{ matrix.model }}" = "bert-99" ] || [ "${{ matrix.model }}" = "bert-99.9" ]; then
47-
category="edge"
48-
else
49-
category="datacenter,edge"
50-
fi
51-
48+
category="datacenter,edge"
5249
if [ -f "gh_action/bin/deactivate" ]; then source gh_action/bin/deactivate; fi
5350
python3 -m venv gh_action
5451
source gh_action/bin/activate

automation/script/docker_utils.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ def prepare_docker_inputs(input_params, docker_settings,
109109
keys += [
110110
"skip_run_cmd", "pre_run_cmds", "run_cmd_prefix", "all_gpus", "num_gpus", "device", "gh_token",
111111
"port_maps", "shm_size", "pass_user_id", "pass_user_group", "extra_run_args", "detached", "interactive",
112-
"dt", "it", "use_host_group_id", "use_host_user_id", "keep_detached", "reuse_existing"
112+
"dt", "it", "use_host_group_id", "use_host_user_id", "keep_detached", "reuse_existing", "use_google_dns"
113113
]
114114
# Collect Dockerfile inputs
115115
docker_inputs = {
@@ -413,7 +413,7 @@ def get_container_path_script(i):
413413
def get_container_path(value, username="mlcuser"):
414414
path_split = value.split(os.sep)
415415
if len(path_split) == 1:
416-
return value
416+
return value, value
417417

418418
new_value = ''
419419
if "cache" in path_split and "local" in path_split:

automation/script/module.py

Lines changed: 2 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -273,31 +273,17 @@ def _run(self, i):
273273
# Useful to set up default flags such as the name of virtual enviroment
274274
extra_cli = os.environ.get('MLC_SCRIPT_EXTRA_CMD', '').strip()
275275
if extra_cli != '':
276-
from cmind import cli
277-
r = cli.parse(extra_cli)
276+
r = convert_args_to_dictionary(extra_cli)
278277
if r['return'] > 0:
279278
return r
280279

281-
mlc_input = r['mlc_input']
280+
mlc_input = r['args_dict']
282281

283282
utils.merge_dicts({'dict1': i,
284283
'dict2': mlc_input,
285284
'append_lists': True,
286285
'append_unique': True})
287286

288-
# Check if has extra tags as a second artifact
289-
# Example: cmr . "_python _tiny"
290-
291-
parsed_artifacts = i.get('parsed_artifacts', [])
292-
if len(parsed_artifacts) > 0:
293-
extra_tags = parsed_artifacts[0][0][0]
294-
if ' ' in extra_tags or ',' in extra_tags:
295-
# Add tags
296-
x = i.get('tags', '')
297-
if x != '':
298-
x += ','
299-
i['tags'] = x + extra_tags.replace(' ', ',')
300-
301287
# Recursion spaces needed to format log and print
302288
recursion_spaces = i.get('recursion_spaces', '')
303289
# Caching selections to avoid asking users again

0 commit comments

Comments
 (0)