File tree Expand file tree Collapse file tree 2 files changed +9
-6
lines changed
app-mlperf-automotive-mlcommons-python
generate-mlperf-inference-user-conf Expand file tree Collapse file tree 2 files changed +9
-6
lines changed Original file line number Diff line number Diff line change @@ -458,7 +458,8 @@ variations:
458
458
bevformer :
459
459
group : models
460
460
default_env :
461
- MLC_MLPERF_MAX_QUERY_COUNT : 6636
461
+ MLC_MLPERF_SINGLESTREAM_TARGET_LATENCY_PERCENTILE : 99.9
462
+ MLC_DEFAULT_MLPERF_MAX_QUERY_COUNT : 6636
462
463
add_deps_recursive :
463
464
pytorch :
464
465
version_max : " 2.5.1"
@@ -495,7 +496,8 @@ variations:
495
496
ssd :
496
497
group : models
497
498
default_env :
498
- MLC_MLPERF_MAX_QUERY_COUNT : 6636
499
+ MLC_MLPERF_SINGLESTREAM_TARGET_LATENCY_PERCENTILE : 99.9
500
+ MLC_DEFAULT_MLPERF_MAX_QUERY_COUNT : 6636
499
501
add_deps_recursive :
500
502
pytorch :
501
503
version_max : " 2.3.1"
@@ -535,7 +537,8 @@ variations:
535
537
deeplabv3plus :
536
538
group : models
537
539
default_env :
538
- MLC_MLPERF_MAX_QUERY_COUNT : 6636
540
+ MLC_MLPERF_SINGLESTREAM_TARGET_LATENCY_PERCENTILE : 99.9
541
+ MLC_DEFAULT_MLPERF_MAX_QUERY_COUNT : 6636
539
542
add_deps_recursive :
540
543
pytorch :
541
544
version_max : " 2.3.1"
Original file line number Diff line number Diff line change @@ -368,11 +368,11 @@ def preprocess(i):
368
368
user_conf += ml_model_name + "." + scenario + ".min_duration = " + \
369
369
env ['MLC_MLPERF_INFERENCE_MIN_DURATION' ] + " \n "
370
370
elif not is_false (env .get ('MLC_MLPERF_USE_MAX_DURATION' , 'yes' )) and not is_true (env .get (
371
- 'MLC_MLPERF_MODEL_EQUAL_ISSUE_MODE' , 'no' )) and env .get ('MLC_BENCHMARK_GROUP ' ) != 'automotive ' :
371
+ 'MLC_MLPERF_MODEL_EQUAL_ISSUE_MODE' , 'no' )) and env .get ('MLC_MLPERF_SINGLESTREAM_TARGET_LATENCY_PERCENTILE ' ) != '99.9 ' :
372
372
user_conf += ml_model_name + "." + scenario + \
373
373
f".max_duration = { max_duration_valid } " + "\n "
374
- if scenario == "SingleStream" and env .get ('MLC_BENCHMARK_GROUP ' ) == 'automotive ' and env .get (
375
- 'MLC_MLPERF_MAX_QUERY_COUNT ' , '' ) != '' and env .get ('MLC_MLPERF_TARGET_LATENCY' , '' ) == '' :
374
+ if scenario == "SingleStream" and env .get ('MLC_MLPERF_SINGLESTREAM_TARGET_LATENCY_PERCENTILE ' ) == '99.9 ' and env .get (
375
+ 'MLC_MLPERF_DEFAULT_MAX_QUERY_COUNT ' , '' ) != '' and env .get ('MLC_MLPERF_TARGET_LATENCY' , '' ) == '' :
376
376
user_conf += ml_model_name + "." + scenario + \
377
377
f".max_query_count = { env .get ('MLC_MLPERF_MAX_QUERY_COUNT' )} " + "\n "
378
378
if scenario == "MultiStream" :
You can’t perform that action at this time.
0 commit comments