@@ -35,30 +35,30 @@ jobs:
35
35
36
36
- name : Test Python venv
37
37
run : |
38
- mlcr --tags= install,python-venv --name=test --quiet
38
+ mlcr install,python-venv --name=test --quiet
39
39
mlc search cache --tags=get,python,virtual,name-test --quiet
40
40
41
41
- name : Test variations
42
42
run : |
43
- mlcr --tags= get,dataset,preprocessed,imagenet,_NHWC --quiet
43
+ mlcr get,dataset,preprocessed,imagenet,_NHWC --quiet
44
44
mlc search cache --tags=get,dataset,preprocessed,imagenet,-_NCHW
45
45
mlc search cache --tags=get,dataset,preprocessed,imagenet,-_NHWC
46
46
47
47
- name : Test versions
48
48
continue-on-error : true
49
49
if : runner.os == 'linux'
50
50
run : |
51
- mlcr --tags= get,generic-python-lib,_package.scipy --version=1.9.3 --quiet
51
+ mlcr get,generic-python-lib,_package.scipy --version=1.9.3 --quiet
52
52
test $? -eq 0 || exit $?
53
- mlcr --tags= get,generic-python-lib,_package.scipy --version=1.9.2 --quiet
53
+ mlcr get,generic-python-lib,_package.scipy --version=1.9.2 --quiet
54
54
test $? -eq 0 || exit $?
55
55
# Need to add find cache here
56
- # mlcr --tags= get,generic-python-lib,_package.scipy --version=1.9.3 --quiet --only_execute_from_cache=True
56
+ # mlcr get,generic-python-lib,_package.scipy --version=1.9.3 --quiet --only_execute_from_cache=True
57
57
# test $? -eq 0 || exit 0
58
58
59
59
- name : Test python install from src
60
60
run : |
61
- mlcr --tags= python,src,install,_shared --version=3.9.10 --quiet
61
+ mlcr python,src,install,_shared --version=3.9.10 --quiet
62
62
mlc search cache --tags=python,src,install,_shared,version-3.9.10
63
63
64
64
test_docker :
@@ -81,11 +81,11 @@ jobs:
81
81
82
82
- name : Run docker container from dockerhub on linux
83
83
run : |
84
- mlcr --tags= run,docker,container --adr.compiler.tags=gcc --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --image_name=cm-script-app-image-classification-onnx-py --env.MLC_DOCKER_RUN_SCRIPT_TAGS=app,image-classification,onnx,python --env.MLC_DOCKER_IMAGE_BASE=ubuntu:22.04 --env.MLC_DOCKER_IMAGE_REPO=cknowledge --quiet
84
+ mlcr run,docker,container --adr.compiler.tags=gcc --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --image_name=cm-script-app-image-classification-onnx-py --env.MLC_DOCKER_RUN_SCRIPT_TAGS=app,image-classification,onnx,python --env.MLC_DOCKER_IMAGE_BASE=ubuntu:22.04 --env.MLC_DOCKER_IMAGE_REPO=cknowledge --quiet
85
85
86
86
- name : Run docker container locally on linux
87
87
run : |
88
- mlcr --tags= run,docker,container --adr.compiler.tags=gcc --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --image_name=mlc-script-app-image-classification-onnx-py --env.MLC_DOCKER_RUN_SCRIPT_TAGS=app,image-classification,onnx,python --env.MLC_DOCKER_IMAGE_BASE=ubuntu:22.04 --env.MLC_DOCKER_IMAGE_REPO=local --quiet
88
+ mlcr run,docker,container --adr.compiler.tags=gcc --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --image_name=mlc-script-app-image-classification-onnx-py --env.MLC_DOCKER_RUN_SCRIPT_TAGS=app,image-classification,onnx,python --env.MLC_DOCKER_IMAGE_BASE=ubuntu:22.04 --env.MLC_DOCKER_IMAGE_REPO=local --quiet
89
89
90
90
test_mlperf_retinanet_cpp_venv :
91
91
runs-on : ubuntu-latest
@@ -107,15 +107,15 @@ jobs:
107
107
108
108
- name : Run MLPerf Inference Retinanet with native and virtual Python
109
109
run : |
110
- mlcr --tags= app,mlperf,inference,generic,_cpp,_retinanet,_onnxruntime,_cpu --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --scenario=Offline --mode=accuracy --test_query_count=10 --rerun --quiet
110
+ mlcr app,mlperf,inference,generic,_cpp,_retinanet,_onnxruntime,_cpu --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --scenario=Offline --mode=accuracy --test_query_count=10 --rerun --quiet
111
111
112
- mlcr --tags= app,mlperf,inference,generic,_cpp,_retinanet,_onnxruntime,_cpu --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --scenario=Offline --mode=performance --test_query_count=10 --rerun --quiet
112
+ mlcr app,mlperf,inference,generic,_cpp,_retinanet,_onnxruntime,_cpu --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --scenario=Offline --mode=performance --test_query_count=10 --rerun --quiet
113
113
114
- mlcr --tags= install,python-venv --version=3.10.8 --name=mlperf --quiet
114
+ mlcr install,python-venv --version=3.10.8 --name=mlperf --quiet
115
115
116
116
export MLC_SCRIPT_EXTRA_CMD="--adr.python.name=mlperf"
117
117
118
- mlcr --tags= run,mlperf,inference,_submission,_short --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --submitter=MLCommons --implementation=cpp --hw_name=default --model=retinanet --backend=onnxruntime --device=cpu --scenario=Offline --quiet
118
+ mlcr run,mlperf,inference,_submission,_short --adr.python.version_min=3.8 --adr.compiler.tags=gcc --adr.openimages-preprocessed.tags=_50 --submitter=MLCommons --implementation=cpp --hw_name=default --model=retinanet --backend=onnxruntime --device=cpu --scenario=Offline --quiet
119
119
120
120
# Step for Linux/MacOS
121
121
- name : Randomly Execute Step (Linux/MacOS)
@@ -160,4 +160,4 @@ jobs:
160
160
git config --global credential.https://github.com.helper "!gh auth git-credential"
161
161
git config --global credential.https://gist.github.com.helper ""
162
162
git config --global credential.https://gist.github.com.helper "!gh auth git-credential"
163
- mlcr --tags= push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_test_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from R50 GH action on ${{ matrix.os }}" --quiet
163
+ mlcr push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_test_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from R50 GH action on ${{ matrix.os }}" --quiet
0 commit comments