Skip to content

Commit 532dd21

Browse files
authored
Support different llama2 variants (#222)
* Strip mlperf inference submitter name * Fix llama2 mlperf variants
1 parent 43fba7d commit 532dd21

File tree

4 files changed

+33
-7
lines changed

4 files changed

+33
-7
lines changed

.github/workflows/test-nvidia-mlperf-inference-implementations.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@ name: MLPerf Inference Nvidia implementations
22

33
on:
44
schedule:
5-
- cron: "05 01 * * *"
5+
- cron: "27 11 * * *"
66

77
jobs:
88
run_nvidia:
@@ -58,7 +58,7 @@ jobs:
5858
pip install --upgrade mlcflow
5959
mlc pull repo mlcommons@mlperf-automations --branch=dev
6060
61-
mlcr --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r5.0-dev --preprocess_submission=yes --pull_changes=yes --pull_inference_changes=yes --execution_mode=valid --gpu_name=$gpu_name --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="GATEOverflow " --hw_name=$hw_name --implementation=nvidia --backend=tensorrt --category=$category --division=closed --docker_dt --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --adr.compiler.tags=gcc --device=cuda --use_model_from_host=yes --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean $docker_string $submission_preprocessor_args --quiet
61+
mlcr --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r5.0-dev --preprocess_submission=yes --pull_changes=yes --pull_inference_changes=yes --execution_mode=valid --gpu_name=$gpu_name --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="GATEOverflow" --hw_name=$hw_name --implementation=nvidia --backend=tensorrt --category=$category --division=closed --docker_dt --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --adr.compiler.tags=gcc --device=cuda --use_model_from_host=yes --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean $docker_string $submission_preprocessor_args --quiet
6262
#mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_unofficial_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from GH action on NVIDIA_$hw_name" --quiet --submission_dir=$HOME/gh_action_submissions --hw_name=$hw_name
6363
mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/GATEOverflow/mlperf_inference_submissions_v5.0 --repo_branch=main --commit_message="Results from GH actions on NVIDIA_$hw_name" --quiet --submission_dir=$HOME/gh_action_submissions --hw_name=$hw_name
6464

script/generate-mlperf-inference-submission/customize.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -150,7 +150,7 @@ def generate_submission(env, state, inp, submission_division):
150150

151151
# Check submitter
152152
if env.get('MLC_MLPERF_SUBMITTER'):
153-
submitter = env['MLC_MLPERF_SUBMITTER']
153+
submitter = env['MLC_MLPERF_SUBMITTER'].strip()
154154
system_meta_tmp['submitter'] = submitter
155155
else:
156156
submitter = system_meta_default['submitter']

script/get-ml-model-llama2/meta.yaml

Lines changed: 26 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -93,23 +93,46 @@ variations:
9393
group: download-source
9494
env:
9595
MLC_DOWNLOAD_SRC: huggingface
96+
70b:
97+
env:
98+
MLC_GIT_CHECKOUT_FOLDER: Llama-2-70b-chat-hf
99+
group: model-size
100+
default: true
101+
default_variations:
102+
huggingface-stub: meta-llama/Llama-2-70b-chat-hf
103+
7b:
104+
env:
105+
MLC_GIT_CHECKOUT_FOLDER: Llama-2-7b-chat-hf
106+
group: model-size
107+
default_variations:
108+
huggingface-stub: meta-llama/Llama-2-7b-chat-hf
109+
110+
70b-fused-qkv:
111+
env:
112+
MLC_GIT_CHECKOUT_FOLDER: Llama-2-70b-fused-qkv-mlperf
113+
group: model-size
114+
96115
meta-llama/Llama-2-70b-chat-hf:
116+
base:
117+
- 70b
97118
adr:
98119
hf-zoo:
99120
tags: _model-stub.meta-llama/Llama-2-70b-chat-hf
100-
default: true
101121
env:
102-
MLC_GIT_CHECKOUT_FOLDER: Llama-2-70b-chat-hf
103122
MLC_MODEL_ZOO_ENV_KEY: LLAMA2
104123
group: huggingface-stub
124+
105125
meta-llama/Llama-2-7b-chat-hf:
126+
base:
127+
- 7b
106128
adr:
107129
hf-zoo:
108130
tags: _model-stub.meta-llama/Llama-2-7b-chat-hf
109131
env:
110-
MLC_GIT_CHECKOUT_FOLDER: Llama-2-7b-chat-hf
111132
MLC_MODEL_ZOO_ENV_KEY: LLAMA2
112133
group: huggingface-stub
134+
135+
113136
nvidia:
114137
default_variations:
115138
framework: pytorch
Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,6 @@
11
rclone config create mlc-llama2 drive config_is_local=false scope=drive.readonly root_folder_id=11tBZvvrh0FCm3XuR5E849K42TqftYdUF
22
rclone config reconnect mlc-llama2:
3-
rclone sync mlc-llama2:${MLC_GIT_CHECKOUT_FOLDER} ${LLAMA2_CHECKPOINT_PATH}/${MLC_GIT_CHECKOUT_FOLDER} -P
3+
cmd="rclone sync mlc-llama2:${MLC_GIT_CHECKOUT_FOLDER} ${LLAMA2_CHECKPOINT_PATH}/${MLC_GIT_CHECKOUT_FOLDER} -P"
4+
echo $cmd
5+
eval $cmd
6+
test $? -eq 0 || exit $?

0 commit comments

Comments
 (0)