@@ -27,9 +27,9 @@ division="closed"
27
27
# run "$MLC_RUN_CMD"
28
28
29
29
POWER=" --power=yes --adr.mlperf-power-client.power_server=192.168.0.15 --adr.mlperf-power-client.port=4950 "
30
- POWER=" "
30
+ POWER=" --env.MLC_GET_PLATFORM_DETAILS=no "
31
31
32
- run " mlcr set,system,performance,mode"
32
+ # run "mlcr set,system,performance,mode"
33
33
34
34
# cpp
35
35
run " mlcr generate-run-cmds,inference,_find-performance \
@@ -51,7 +51,7 @@ run "mlcr generate-run-cmds,inference,_submission \
51
51
--execution_mode=valid \
52
52
--skip_submission_generation=yes \
53
53
${POWER} \
54
- --results_dir= $HOME /results_dir "
54
+ "
55
55
56
56
run " mlcr generate-run-cmds,inference,_submission \
57
57
--model=retinanet --implementation=cpp --device=cpu --backend=onnxruntime \
@@ -61,7 +61,7 @@ run "mlcr generate-run-cmds,inference,_submission \
61
61
--execution_mode=valid \
62
62
--skip_submission_generation=yes \
63
63
${POWER} \
64
- --results_dir= $HOME /results_dir "
64
+ "
65
65
66
66
run " mlcr generate-run-cmds,inference,_submission \
67
67
--model=resnet50 --implementation=cpp --device=cpu --backend=onnxruntime \
@@ -71,7 +71,7 @@ run "mlcr generate-run-cmds,inference,_submission \
71
71
--execution_mode=valid \
72
72
--skip_submission_generation=yes \
73
73
${POWER} \
74
- --results_dir= $HOME /results_dir "
74
+ "
75
75
76
76
run " mlcr generate-run-cmds,inference,_submission \
77
77
--model=retinanet --implementation=cpp --device=cpu --backend=onnxruntime \
@@ -81,7 +81,7 @@ run "mlcr generate-run-cmds,inference,_submission \
81
81
--execution_mode=valid \
82
82
--skip_submission_generation=yes \
83
83
${POWER} \
84
- --results_dir= $HOME /results_dir "
84
+ "
85
85
86
86
# GPU
87
87
@@ -106,7 +106,7 @@ run "mlcr generate-run-cmds,inference,_submission \
106
106
--execution_mode=valid \
107
107
--skip_submission_generation=yes \
108
108
${POWER} \
109
- --results_dir= $HOME /results_dir "
109
+ "
110
110
111
111
run " mlcr generate-run-cmds,inference,_submission \
112
112
--model=retinanet --implementation=cpp --device=cuda --backend=onnxruntime \
@@ -116,7 +116,7 @@ run "mlcr generate-run-cmds,inference,_submission \
116
116
--execution_mode=valid \
117
117
--skip_submission_generation=yes \
118
118
${POWER} \
119
- --results_dir= $HOME /results_dir "
119
+ "
120
120
121
121
122
122
run " mlcr generate-run-cmds,inference,_submission \
@@ -128,7 +128,7 @@ run "mlcr generate-run-cmds,inference,_submission \
128
128
--execution_mode=valid \
129
129
--skip_submission_generation=yes \
130
130
${POWER} \
131
- --results_dir= $HOME /results_dir "
131
+ "
132
132
133
133
run " mlcr generate-run-cmds,inference,_submission \
134
134
--model=retinanet --implementation=cpp --device=cuda --backend=onnxruntime \
@@ -138,7 +138,7 @@ run "mlcr generate-run-cmds,inference,_submission \
138
138
--execution_mode=valid \
139
139
--skip_submission_generation=yes \
140
140
${POWER} \
141
- --results_dir= $HOME /results_dir "
141
+ "
142
142
143
143
# multistream
144
144
run " mlcr generate-run-cmds,inference,_submission \
@@ -150,7 +150,7 @@ run "mlcr generate-run-cmds,inference,_submission \
150
150
--execution_mode=valid \
151
151
--skip_submission_generation=yes \
152
152
${POWER} \
153
- --results_dir= $HOME /results_dir "
153
+ "
154
154
155
155
run " mlcr generate-run-cmds,inference,_submission \
156
156
--model=retinanet --implementation=cpp --device=cuda --backend=onnxruntime \
@@ -160,4 +160,4 @@ run "mlcr generate-run-cmds,inference,_submission \
160
160
--execution_mode=valid \
161
161
--skip_submission_generation=yes \
162
162
${POWER} \
163
- --results_dir= $HOME /results_dir "
163
+ "
0 commit comments