Skip to content

Commit 092494a

Browse files
committed
Auto-merge updates from auto-update branch
2 parents ff6a2b9 + b45598e commit 092494a

File tree

11 files changed

+250
-218
lines changed

11 files changed

+250
-218
lines changed

open/MLCommons/measurements/default-mlcommons_cpp-cpu-onnxruntime-default_config/retinanet/offline/README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -16,16 +16,16 @@ pip install -U mlcflow
1616

1717
mlc rm cache -f
1818

19-
mlc pull repo anandhu-eng@mlperf-automations --checkout=f5f7996026c2b55a25721b7681ab386e03b1aa17
19+
mlc pull repo GATEOverflow@mlperf-automations --checkout=a1a2fc86e398d44813bb04231ad36addf3621ea6
2020

2121

2222
```
2323
*Note that if you want to use the [latest automation recipes](https://docs.mlcommons.org/inference) for MLPerf,
24-
you should simply reload anandhu-eng@mlperf-automations without checkout and clean MLC cache as follows:*
24+
you should simply reload GATEOverflow@mlperf-automations without checkout and clean MLC cache as follows:*
2525

2626
```bash
27-
mlc rm repo anandhu-eng@mlperf-automations
28-
mlc pull repo anandhu-eng@mlperf-automations
27+
mlc rm repo GATEOverflow@mlperf-automations
28+
mlc pull repo GATEOverflow@mlperf-automations
2929
mlc rm cache -f
3030

3131
```
@@ -40,4 +40,4 @@ Model Precision: fp32
4040
`mAP`: `49.593`, Required accuracy for closed division `>= 37.1745`
4141

4242
### Performance Results
43-
`Samples per second`: `0.420673`
43+
`Samples per second`: `0.421428`

open/MLCommons/measurements/default-mlcommons_cpp-cpu-onnxruntime-default_config/retinanet/offline/accuracy_console.out

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
1-
User Conf path: /home/runner/MLC/repos/anandhu-eng@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/3cd3871c7c814de0944bb7f496f872c5.conf
2-
Dataset Preprocessed path: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_77700f64
3-
Dataset List filepath: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_77700f64/annotations/openimages-mlperf.json
1+
User Conf path: /home/runner/MLC/repos/GATEOverflow@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/03bc05c344504674bf130a4c1f3a8d49.conf
2+
Dataset Preprocessed path: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_b6f591b8
3+
Dataset List filepath: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_b6f591b8/annotations/openimages-mlperf.json
44
Scenario: Offline
55
Mode: AccuracyOnly
66
Batch size: 1

open/MLCommons/measurements/default-mlcommons_cpp-cpu-onnxruntime-default_config/retinanet/offline/cpu_info.json

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,6 @@
2020
"MLC_HOST_CPU_L2_CACHE_SIZE": "1 MiB (2 instances)",
2121
"MLC_HOST_CPU_L3_CACHE_SIZE": "32 MiB (1 instance)",
2222
"MLC_HOST_CPU_NUMA_NODES": "1",
23-
"MLC_HOST_CPU_START_CORES": "0",
2423
"MLC_HOST_CPU_TOTAL_LOGICAL_CORES": "4",
2524
"MLC_HOST_CPU_TOTAL_PHYSICAL_CORES": "2",
2625
"MLC_HOST_CPU_PHYSICAL_CORES_LIST": "0-1",

open/MLCommons/measurements/default-mlcommons_cpp-cpu-onnxruntime-default_config/retinanet/offline/mlc-deps.mmd

Lines changed: 17 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -2,13 +2,16 @@ graph TD
22
app-mlperf-inference,d775cac873ee4231_(_cpp,_retinanet,_onnxruntime,_cpu,_test,_r5.1-dev_default,_offline_) --> detect,os
33
app-mlperf-inference,d775cac873ee4231_(_cpp,_retinanet,_onnxruntime,_cpu,_test,_r5.1-dev_default,_offline_) --> get,sys-utils-cm
44
app-mlperf-inference,d775cac873ee4231_(_cpp,_retinanet,_onnxruntime,_cpu,_test,_r5.1-dev_default,_offline_) --> get,python
5-
app-mlperf-inference,d775cac873ee4231_(_cpp,_retinanet,_onnxruntime,_cpu,_test,_r5.1-dev_default,_offline_) --> get,mlcommons,inference,src
6-
get-mlperf-inference-utils,e341e5f86d8342e5 --> get,mlperf,inference,src
5+
get-mlperf-inference-src,4b57186581024797_(_branch.master_) --> detect,os
6+
get-mlperf-inference-src,4b57186581024797_(_branch.master_) --> get,python3
7+
get-mlperf-inference-src,4b57186581024797_(_branch.master_) --> get,git,repo,_branch.master,_repo.https://github.com/mlcommons/inference
8+
app-mlperf-inference,d775cac873ee4231_(_cpp,_retinanet,_onnxruntime,_cpu,_test,_r5.1-dev_default,_offline_) --> get,mlcommons,inference,src,_branch.master
9+
get-mlperf-inference-utils,e341e5f86d8342e5 --> get,mlperf,inference,src,_branch.master
710
app-mlperf-inference,d775cac873ee4231_(_cpp,_retinanet,_onnxruntime,_cpu,_test,_r5.1-dev_default,_offline_) --> get,mlperf,inference,utils
8-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> detect,os
11+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> detect,os
912
detect-cpu,586c8a43320142f7 --> detect,os
10-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> detect,cpu
11-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> get,sys-utils-cm
13+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> detect,cpu
14+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> get,sys-utils-cm
1215
get-mlperf-inference-loadgen,64c3d98d0ba04950_(_wg-inference_) --> detect,os
1316
get-mlperf-inference-loadgen,64c3d98d0ba04950_(_wg-inference_) --> get,python3
1417
get-mlperf-inference-loadgen,64c3d98d0ba04950_(_wg-inference_) --> get,mlcommons,inference,src
@@ -40,27 +43,27 @@ graph TD
4043
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
4144
get-generic-python-lib,94b62a682bc44791_(_package.setuptools_) --> get,generic-python-lib,_pip
4245
get-mlperf-inference-loadgen,64c3d98d0ba04950_(_wg-inference_) --> get,generic-python-lib,_package.setuptools
43-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> get,loadgen,_wg-inference
44-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> get,mlcommons,inference,src
45-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> get,lib,onnxruntime,lang-cpp,_cpu
46-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> get,dataset,preprocessed,openimages,_validation,_NCHW,_50
47-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> get,ml-model,retinanet,_onnx,_fp32
46+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> get,loadgen,_wg-inference
47+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> get,mlcommons,inference,src,_branch.master
48+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> get,lib,onnxruntime,lang-cpp,_cpu
49+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> get,dataset,preprocessed,openimages,_validation,_NCHW,_50
50+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> get,ml-model,retinanet,_onnx,_fp32
4851
generate-mlperf-inference-user-conf,3af4475745964b93_(_wg-inference_) --> detect,os
4952
detect-cpu,586c8a43320142f7 --> detect,os
5053
generate-mlperf-inference-user-conf,3af4475745964b93_(_wg-inference_) --> detect,cpu
5154
generate-mlperf-inference-user-conf,3af4475745964b93_(_wg-inference_) --> get,python
5255
get-mlperf-inference-sut-configs,c2fbf72009e2445b --> get,cache,dir,_name.mlperf-inference-sut-configs
5356
generate-mlperf-inference-user-conf,3af4475745964b93_(_wg-inference_) --> get,sut,configs
54-
generate-mlperf-inference-user-conf,3af4475745964b93_(_wg-inference_) --> get,mlcommons,inference,src
55-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> generate,user-conf,mlperf,inference,_wg-inference
57+
generate-mlperf-inference-user-conf,3af4475745964b93_(_wg-inference_) --> get,mlcommons,inference,src,_branch.master
58+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> generate,user-conf,mlperf,inference,_wg-inference
5659
detect-cpu,586c8a43320142f7 --> detect,os
5760
compile-program,c05042ba005a4bfa --> detect,cpu
5861
compile-program,c05042ba005a4bfa --> get,compiler,gcc
5962
detect-cpu,586c8a43320142f7 --> detect,os
6063
get-compiler-flags,31be8b74a69742f8 --> detect,cpu
6164
compile-program,c05042ba005a4bfa --> get,compiler-flags
62-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> compile,cpp-program
65+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> compile,cpp-program
6366
detect-cpu,586c8a43320142f7 --> detect,os
6467
benchmark-program,19f369ef47084895 --> detect,cpu
6568
benchmark-program-mlperf,cfff0132a8aa4018 --> benchmark-program,program
66-
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_offline,_cpu,_onnxruntime,_retinanet_) --> benchmark-mlperf
69+
app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf_(_retinanet,_onnxruntime,_cpu,_offline_) --> benchmark-mlperf

open/MLCommons/measurements/default-mlcommons_cpp-cpu-onnxruntime-default_config/retinanet/offline/mlc-version-info.json

Lines changed: 53 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -31,22 +31,52 @@
3131
}
3232
},
3333
{
34-
"get,mlcommons,inference,src": {
34+
"detect,os": {
35+
"script_uid": "863735b7db8c44fc",
36+
"script_alias": "detect-os",
37+
"script_tags": "detect-os,detect,os,info",
38+
"script_variations": "",
39+
"version": "",
40+
"parent": "get-mlperf-inference-src,4b57186581024797 ( branch.master )"
41+
}
42+
},
43+
{
44+
"get,python3": {
45+
"script_uid": "d0b5dd74373f4a62",
46+
"script_alias": "get-python3",
47+
"script_tags": "get,python,python3,get-python,get-python3",
48+
"script_variations": "",
49+
"version": "3.10.8",
50+
"parent": "get-mlperf-inference-src,4b57186581024797 ( branch.master )"
51+
}
52+
},
53+
{
54+
"get,git,repo,_branch.master,_repo.https://github.com/mlcommons/inference": {
55+
"script_uid": "ed603e7292974f10",
56+
"script_alias": "get-git-repo",
57+
"script_tags": "get,git,repo,repository,clone",
58+
"script_variations": "branch.master,repo.https://github.com/mlcommons/inference",
59+
"version": "",
60+
"parent": "get-mlperf-inference-src,4b57186581024797 ( branch.master )"
61+
}
62+
},
63+
{
64+
"get,mlcommons,inference,src,_branch.master": {
3565
"script_uid": "4b57186581024797",
3666
"script_alias": "get-mlperf-inference-src",
3767
"script_tags": "get,src,source,inference,inference-src,inference-source,mlperf,mlcommons",
38-
"script_variations": "",
39-
"version": "master-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
68+
"script_variations": "branch.master",
69+
"version": "custom-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
4070
"parent": "app-mlperf-inference,d775cac873ee4231 ( cpp,_retinanet,_onnxruntime,_cpu,_test,_r5.1-dev_default,_offline )"
4171
}
4272
},
4373
{
44-
"get,mlperf,inference,src": {
74+
"get,mlperf,inference,src,_branch.master": {
4575
"script_uid": "4b57186581024797",
4676
"script_alias": "get-mlperf-inference-src",
4777
"script_tags": "get,src,source,inference,inference-src,inference-source,mlperf,mlcommons",
48-
"script_variations": "",
49-
"version": "master-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
78+
"script_variations": "branch.master",
79+
"version": "custom-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
5080
"parent": "get-mlperf-inference-utils,e341e5f86d8342e5"
5181
}
5282
},
@@ -67,7 +97,7 @@
6797
"script_tags": "detect-os,detect,os,info",
6898
"script_variations": "",
6999
"version": "",
70-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
100+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
71101
}
72102
},
73103
{
@@ -87,7 +117,7 @@
87117
"script_tags": "detect,cpu,detect-cpu,info",
88118
"script_variations": "",
89119
"version": "",
90-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
120+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
91121
}
92122
},
93123
{
@@ -97,7 +127,7 @@
97127
"script_tags": "get,sys-utils-cm,sys-utils-mlc",
98128
"script_variations": "",
99129
"version": "",
100-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
130+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
101131
}
102132
},
103133
{
@@ -417,17 +447,17 @@
417447
"script_tags": "get,loadgen,inference,inference-loadgen,mlperf,mlcommons",
418448
"script_variations": "wg-inference",
419449
"version": "master",
420-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
450+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
421451
}
422452
},
423453
{
424-
"get,mlcommons,inference,src": {
454+
"get,mlcommons,inference,src,_branch.master": {
425455
"script_uid": "4b57186581024797",
426456
"script_alias": "get-mlperf-inference-src",
427457
"script_tags": "get,src,source,inference,inference-src,inference-source,mlperf,mlcommons",
428-
"script_variations": "",
429-
"version": "master-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
430-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
458+
"script_variations": "branch.master",
459+
"version": "custom-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
460+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
431461
}
432462
},
433463
{
@@ -437,7 +467,7 @@
437467
"script_tags": "install,onnxruntime,get,prebuilt,lib,lang-c,lang-cpp",
438468
"script_variations": "cpu",
439469
"version": "",
440-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
470+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
441471
}
442472
},
443473
{
@@ -447,7 +477,7 @@
447477
"script_tags": "get,dataset,openimages,open-images,object-detection,preprocessed",
448478
"script_variations": "validation,NCHW,50",
449479
"version": "",
450-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
480+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
451481
}
452482
},
453483
{
@@ -457,7 +487,7 @@
457487
"script_tags": "get,ml-model,raw,resnext50,retinanet,object-detection",
458488
"script_variations": "onnx,fp32",
459489
"version": "",
460-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
490+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
461491
}
462492
},
463493
{
@@ -521,12 +551,12 @@
521551
}
522552
},
523553
{
524-
"get,mlcommons,inference,src": {
554+
"get,mlcommons,inference,src,_branch.master": {
525555
"script_uid": "4b57186581024797",
526556
"script_alias": "get-mlperf-inference-src",
527557
"script_tags": "get,src,source,inference,inference-src,inference-source,mlperf,mlcommons",
528-
"script_variations": "",
529-
"version": "master-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
558+
"script_variations": "branch.master",
559+
"version": "custom-git-0a3570efb0309b5581f2831d84c05fe5483b5ef7",
530560
"parent": "generate-mlperf-inference-user-conf,3af4475745964b93 ( wg-inference )"
531561
}
532562
},
@@ -537,7 +567,7 @@
537567
"script_tags": "generate,mlperf,inference,user-conf,inference-user-conf",
538568
"script_variations": "wg-inference",
539569
"version": "",
540-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
570+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
541571
}
542572
},
543573
{
@@ -607,7 +637,7 @@
607637
"script_tags": "compile,program,c-program,cpp-program,compile-program,compile-c-program,compile-cpp-program",
608638
"script_variations": "",
609639
"version": "",
610-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
640+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
611641
}
612642
},
613643
{
@@ -647,7 +677,7 @@
647677
"script_tags": "mlperf,benchmark-mlperf",
648678
"script_variations": "",
649679
"version": "",
650-
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( offline,_cpu,_onnxruntime,_retinanet )"
680+
"parent": "app-mlperf-inference-mlcommons-cpp,bf62405e6c7a44bf ( retinanet,_onnxruntime,_cpu,_offline )"
651681
}
652682
}
653683
]

open/MLCommons/measurements/default-mlcommons_cpp-cpu-onnxruntime-default_config/retinanet/offline/os_info.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,6 @@
2828
"MLC_HOST_PYTHON_BITS": "64",
2929
"MLC_HOST_SYSTEM_NAME": "pkrvmbietmlfzoi",
3030
"+PATH": [
31-
"/home/runner/MLC/repos/local/cache/install-python-src_5d87d5fe/install/bin"
31+
"/home/runner/MLC/repos/local/cache/install-python-src_bb463f18/install/bin"
3232
]
3333
}

open/MLCommons/measurements/default-mlcommons_cpp-cpu-onnxruntime-default_config/retinanet/offline/performance_console.out

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
1-
User Conf path: /home/runner/MLC/repos/anandhu-eng@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/a5eda5c064ea455197fa6898d7d4c70c.conf
2-
Dataset Preprocessed path: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_77700f64
3-
Dataset List filepath: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_77700f64/annotations/openimages-mlperf.json
1+
User Conf path: /home/runner/MLC/repos/GATEOverflow@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/fd4821322c6d4bfa80f87671fa0c9390.conf
2+
Dataset Preprocessed path: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_b6f591b8
3+
Dataset List filepath: /home/runner/MLC/repos/local/cache/get-preprocessed-dataset-openimages_b6f591b8/annotations/openimages-mlperf.json
44
Scenario: Offline
55
Mode: PerformanceOnly
66
Batch size: 1

0 commit comments

Comments
 (0)