diff --git a/.github/workflows/test-amd-mlperf-inference-implementations.yml b/.github/workflows/test-amd-mlperf-inference-implementations.yml index 9ff0b4da4..4c4b6f749 100644 --- a/.github/workflows/test-amd-mlperf-inference-implementations.yml +++ b/.github/workflows/test-amd-mlperf-inference-implementations.yml @@ -20,7 +20,7 @@ jobs: python3 -m venv gh_action source gh_action/bin/activate export MLC_REPOS=$HOME/GH_MLC - pip install --upgrade cm4mlops - cm pull repo + pip install --upgrade mlc-scripts + mlc pull repo mlcr --tags=run-mlperf,inference,_all-scenarios,_full,_r4.1-dev --execution_mode=valid --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=IntelSPR.24c --implementation=amd --backend=pytorch --category=datacenter --division=open --scenario=Offline --docker_dt=yes --docker_it=no --docker_mlc_repo=gateoverflow@mlperf-automations --docker_mlc_repo_branch=dev --adr.compiler.tags=gcc --device=rocm --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean --docker --quiet --docker_skip_run_cmd=yes # mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_unofficial_submissions_v5.0 --repo_branch=dev --commit_message="Results from GH action on SPR.24c" --quiet --submission_dir=$HOME/gh_action_submissions --hw_name=IntelSPR.24c diff --git a/.github/workflows/test-intel-mlperf-inference-implementations.yml b/.github/workflows/test-intel-mlperf-inference-implementations.yml index d56302111..0041f9762 100644 --- a/.github/workflows/test-intel-mlperf-inference-implementations.yml +++ b/.github/workflows/test-intel-mlperf-inference-implementations.yml @@ -20,7 +20,7 @@ jobs: python3 -m venv gh_action_conda source gh_action_conda/bin/activate export MLC_REPOS=$HOME/GH_MLC - pip install --upgrade cm4mlops + pip install --upgrade mlc-scripts pip install tabulate mlcr --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r4.1-dev --preprocess_submission=yes --execution_mode=valid --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=IntelSPR.24c --implementation=intel --backend=pytorch --category=datacenter --division=open --scenario=Offline --docker_dt=yes --docker_it=no --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --adr.compiler.tags=gcc --device=cpu --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean --docker --quiet mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_unofficial_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from GH action on SPR.24c" --quiet --submission_dir=$HOME/gh_action_submissions --hw_name=IntelSPR.24c diff --git a/.github/workflows/test-mlperf-wheel.yml b/.github/workflows/test-mlcscripts-wheel.yml similarity index 83% rename from .github/workflows/test-mlperf-wheel.yml rename to .github/workflows/test-mlcscripts-wheel.yml index 2ff1595fb..7abf40bbc 100644 --- a/.github/workflows/test-mlperf-wheel.yml +++ b/.github/workflows/test-mlcscripts-wheel.yml @@ -1,4 +1,4 @@ -name: Build Python Wheel +name: Build mlc-scripts Wheel on: pull_request: @@ -6,7 +6,7 @@ on: - main - dev paths: - - '.github/workflows/test-mlperf-wheel.yml' + - '.github/workflows/test-mlcscripts-wheel.yml' - 'setup.py' jobs: @@ -16,6 +16,9 @@ jobs: matrix: os: [macos-latest, ubuntu-latest, windows-latest] python-version: [ '3.8', '3.13'] + exclude: + - os: windows-latest + python-version: "3.8" runs-on: ${{ matrix.os }} diff --git a/.github/workflows/test-mlperf-inference-gptj.yml b/.github/workflows/test-mlperf-inference-gptj.yml index c99c503ff..341e2e818 100644 --- a/.github/workflows/test-mlperf-inference-gptj.yml +++ b/.github/workflows/test-mlperf-inference-gptj.yml @@ -24,8 +24,8 @@ jobs: source gh_action/bin/deactivate || python3 -m venv gh_action source gh_action/bin/activate export MLC_REPOS=$HOME/GH_MLC - python3 -m pip install cm4mlops - cm pull repo + python3 -m pip install --upgrade mlc-scripts + mlc pull repo mlcr --tags=run-mlperf,inference,_submission,_short --submitter="MLCommons" --docker --pull_changes=yes --pull_inference_changes=yes --model=gptj-99 --backend=${{ matrix.backend }} --device=cuda --scenario=Offline --test_query_count=1 --precision=${{ matrix.precision }} --target_qps=1 --quiet --docker_it=no --docker_mlc_repo=gateoverflow@mlperf-automations --docker_mlc_repo_branch=dev --adr.compiler.tags=gcc --beam_size=1 --hw_name=gh_action --docker_dt=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --get_platform_details=yes --implementation=reference --clean mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_test_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from self hosted Github actions - NVIDIARTX4090" --quiet --submission_dir=$HOME/gh_action_submissions diff --git a/.github/workflows/test-mlperf-inference-llama2.yml b/.github/workflows/test-mlperf-inference-llama2.yml index 8de010505..70e4e4909 100644 --- a/.github/workflows/test-mlperf-inference-llama2.yml +++ b/.github/workflows/test-mlperf-inference-llama2.yml @@ -25,9 +25,9 @@ jobs: source gh_action/bin/deactivate || python3 -m venv gh_action source gh_action/bin/activate export MLC_REPOS=$HOME/GH_MLC - pip install cm4mlops + pip install mlc-scripts pip install tabulate - cm pull repo + mlc pull repo pip install "huggingface_hub[cli]" git config --global credential.helper store huggingface-cli login --token ${{ secrets.HF_TOKEN }} --add-to-git-credential diff --git a/.github/workflows/test-mlperf-inference-mixtral.yml b/.github/workflows/test-mlperf-inference-mixtral.yml index 26b369c09..e48cbb1e9 100644 --- a/.github/workflows/test-mlperf-inference-mixtral.yml +++ b/.github/workflows/test-mlperf-inference-mixtral.yml @@ -26,10 +26,10 @@ jobs: source gh_action/bin/deactivate || python3 -m venv gh_action source gh_action/bin/activate export MLC_REPOS=$HOME/GH_MLC - pip install cm4mlops + pip install --upgrade mlc-scripts pip install "huggingface_hub[cli]" git config --global credential.helper store huggingface-cli login --token ${{ secrets.HF_TOKEN }} --add-to-git-credential - cm pull repo + mlc pull repo mlcr --tags=run-mlperf,inference,_submission,_short --adr.inference-src.tags=_branch.dev --submitter="MLCommons" --pull_changes=yes --pull_inference_changes=yes --model=mixtral-8x7b --implementation=reference --batch_size=1 --precision=${{ matrix.precision }} --backend=${{ matrix.backend }} --category=datacenter --scenario=Offline --execution_mode=test --device=${{ matrix.device }} --docker_it=no --docker_mlc_repo=gateoverflow@mlperf-automations --adr.compiler.tags=gcc --hw_name=gh_action --docker_dt=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --docker --quiet --test_query_count=3 --target_qps=0.001 --clean --env.MLC_MLPERF_MODEL_MIXTRAL_8X7B_DOWNLOAD_TO_HOST=yes --env.MLC_MLPERF_DATASET_MIXTRAL_8X7B_DOWNLOAD_TO_HOST=yes --adr.openorca-mbxp-gsm8k-combined-preprocessed.tags=_size.1 mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_test_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from self hosted Github actions - GO-phoenix" --quiet --submission_dir=$HOME/gh_action_submissions diff --git a/.github/workflows/test-mlperf-inference-resnet50.yml b/.github/workflows/test-mlperf-inference-resnet50.yml index 85bcd3cc2..4b7413d97 100644 --- a/.github/workflows/test-mlperf-inference-resnet50.yml +++ b/.github/workflows/test-mlperf-inference-resnet50.yml @@ -61,6 +61,15 @@ jobs: - name: Randomly Execute Step id: random-check run: | + if [[ "$RUNNER_OS" == "Windows" ]]; then + $RANDOM_NUMBER = Get-Random -Maximum 10 + Write-Host "Random number is $RANDOM_NUMBER" + if ($RANDOM_NUMBER -eq 0) { + Write-Host "run_step=true" | Out-File -FilePath $Env:GITHUB_ENV -Append + } else { + Write-Host "run_step=false" | Out-File -FilePath $Env:GITHUB_ENV -Append + } + else RANDOM_NUMBER=$((RANDOM % 10)) echo "Random number is $RANDOM_NUMBER" if [ "$RANDOM_NUMBER" -eq 0 ]; then @@ -68,6 +77,8 @@ jobs: else echo "run_step=false" >> $GITHUB_ENV fi + fi + - name: Retrieve secrets from Keeper if: github.repository_owner == 'mlcommons' && env.run_step == 'true' id: ksecrets diff --git a/.github/workflows/test-mlperf-inference-rnnt.yml b/.github/workflows/test-mlperf-inference-rnnt.yml index 352272370..89ec6e4e2 100644 --- a/.github/workflows/test-mlperf-inference-rnnt.yml +++ b/.github/workflows/test-mlperf-inference-rnnt.yml @@ -30,10 +30,10 @@ jobs: python-version: ${{ matrix.python-version }} - name: Install dependencies on Unix Platforms run: | - MLC_PULL_DEFAULT_MLOPS_REPO=no pip install cm4mlops + pip install mlcflow - name: Pull MLOps repository run: | - cm pull repo --url=${{ github.event.pull_request.head.repo.html_url }} --checkout=${{ github.event.pull_request.head.ref }} + mlc pull repo --url=${{ github.event.pull_request.head.repo.html_url }} --checkout=${{ github.event.pull_request.head.ref }} mlcr --quiet --tags=get,sys-utils-cm - name: Test MLPerf Inference RNNT run: | diff --git a/.github/workflows/test-mlperf-inference-sdxl.yaml b/.github/workflows/test-mlperf-inference-sdxl.yaml index 7f7ce1fea..2e287a0be 100644 --- a/.github/workflows/test-mlperf-inference-sdxl.yaml +++ b/.github/workflows/test-mlperf-inference-sdxl.yaml @@ -19,7 +19,7 @@ jobs: source gh_action/bin/deactivate || python3 -m venv gh_action source gh_action/bin/activate export MLC_REPOS=$HOME/GH_MLC - python3 -m pip install cm4mlops - cm pull repo + python3 -m pip install mlc-scripts + mlc pull repo mlcr --tags=run-mlperf,inference,_submission,_short --submitter="MLCommons" --pull_changes=yes --pull_inference_changes=yes --docker --model=sdxl --backend=${{ matrix.backend }} --device=cuda --scenario=Offline --test_query_count=1 --precision=${{ matrix.precision }} --quiet --docker_it=no --docker_mlc_repo=gateoverflow@mlperf-automations --adr.compiler.tags=gcc --hw_name=gh_action --docker_dt=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --env.MLC_MLPERF_MODEL_SDXL_DOWNLOAD_TO_HOST=yes --clean mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_test_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from self hosted Github actions - NVIDIARTX4090" --quiet --submission_dir=$HOME/gh_action_submissions diff --git a/.github/workflows/test-nvidia-mlperf-inference-implementations.yml b/.github/workflows/test-nvidia-mlperf-inference-implementations.yml index 86f06873d..30f74e605 100644 --- a/.github/workflows/test-nvidia-mlperf-inference-implementations.yml +++ b/.github/workflows/test-nvidia-mlperf-inference-implementations.yml @@ -2,7 +2,7 @@ name: MLPerf Inference Nvidia implementations on: schedule: - - cron: "58 23 * * *" #to be adjusted + - cron: "45 17 * * *" #to be adjusted jobs: run_nvidia: @@ -17,7 +17,7 @@ jobs: strategy: fail-fast: false matrix: - system: [ "GO-spr", "phoenix-Amd-Am5", "GO-i9"] + system: [ "GO-spr", "phoenix", "GO-i9"] # system: [ "mlc-server" ] python-version: [ "3.12" ] model: [ "resnet50", "retinanet", "bert-99", "bert-99.9", "gptj-99.9", "3d-unet-99.9", "sdxl" ] @@ -48,9 +48,9 @@ jobs: python3 -m venv gh_action source gh_action/bin/activate export MLC_REPOS=$HOME/GH_MLC - MLC_PULL_DEFAULT_MLOPS_REPO=no pip install --upgrade cm4mlops - cm pull repo + pip install --upgrade mlcflow + mlc pull repo mlcommons@mlperf-automations --branch=dev - mlcr --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r4.1-dev --preprocess_submission=yes --pull_changes=yes --pull_inference_changes=yes --execution_mode=valid --gpu_name=$gpu_name --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=$hw_name --implementation=nvidia --backend=tensorrt --category=datacenter,edge --division=closed --docker_dt=yes --docker_it=no --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --adr.compiler.tags=gcc --device=cuda --use_model_from_host=yes --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean $docker_string --quiet + mlcr --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r5.0-dev --preprocess_submission=yes --pull_changes=yes --pull_inference_changes=yes --execution_mode=valid --gpu_name=$gpu_name --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=$hw_name --implementation=nvidia --backend=tensorrt --category=datacenter,edge --division=closed --docker_dt --docker_mlc_repo=mlcommons@mlperf-automations --docker_mlc_repo_branch=dev --adr.compiler.tags=gcc --device=cuda --use_model_from_host=yes --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean $docker_string --quiet mlcr --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/mlcommons/mlperf_inference_unofficial_submissions_v5.0 --repo_branch=auto-update --commit_message="Results from GH action on NVIDIA_$hw_name" --quiet --submission_dir=$HOME/gh_action_submissions --hw_name=$hw_name diff --git a/.github/workflows/test-qaic-compute-sdk-build.yml b/.github/workflows/test-qaic-compute-sdk-build.yml index 810da9e8e..0dff27cd0 100644 --- a/.github/workflows/test-qaic-compute-sdk-build.yml +++ b/.github/workflows/test-qaic-compute-sdk-build.yml @@ -26,7 +26,7 @@ jobs: python-version: ${{ matrix.python-version }} - name: Install dependencies run: | - MLC_PULL_DEFAULT_MLOPS_REPO=no pip install cm4mlops + pip install mlc-scripts mlcr --tags=get,sys-utils-cm --quiet - name: Test QAIC Compute SDK for compilation diff --git a/.github/workflows/test-qaic-software-kit.yml b/.github/workflows/test-qaic-software-kit.yml index 127de2323..5cbfc0add 100644 --- a/.github/workflows/test-qaic-software-kit.yml +++ b/.github/workflows/test-qaic-software-kit.yml @@ -31,7 +31,7 @@ jobs: python-version: ${{ matrix.python-version }} - name: Pull MLOps repository run: | - pip install mlperf + pip install mlc-scripts mlcr --tags=get,sys-utils-mlc --quiet - name: Test Software Kit for compilation on Ubuntu 20.04 diff --git a/automation/script/docker_utils.py b/automation/script/docker_utils.py index 6379c515f..f67211fcb 100644 --- a/automation/script/docker_utils.py +++ b/automation/script/docker_utils.py @@ -60,7 +60,7 @@ def process_mounts(mounts, env, docker_settings, f_run_cmd): for placeholder in container_placeholders: if placeholder in env: new_container_mount, container_env_key = get_container_path( - env[placeholder]) + env[placeholder], docker_settings.get('user', 'mlcuser')) else: # Skip mount if variable is missing mounts[index] = None break @@ -400,18 +400,19 @@ def get_host_path(value): def get_container_path_script(i): tmp_dep_cached_path = i['tmp_dep_cached_path'] - value_mnt, value_env = get_container_path(tmp_dep_cached_path) + value_mnt, value_env = get_container_path( + tmp_dep_cached_path, os.getlogin()) return {'return': 0, 'value_mnt': value_mnt, 'value_env': value_env} -def get_container_path(value): +def get_container_path(value, username="mlcuser"): path_split = value.split(os.sep) if len(path_split) == 1: return value new_value = '' if "cache" in path_split and "local" in path_split: - new_path_split = ["", "home", "mlcuser", "MLC", "repos"] + new_path_split = ["", "home", username, "MLC", "repos"] repo_entry_index = path_split.index("local") if len(path_split) >= repo_entry_index + 3: new_path_split1 = new_path_split + \ diff --git a/pyproject.toml b/pyproject.toml index d96eee0c6..7c1b4e8a1 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -4,6 +4,7 @@ build-backend = "setuptools.build_meta" [project] name = "mlc-scripts" +version="0.0.5" description = "Automation scripts for running ML applications using MLC interface" authors = [ { name = "MLCommons", email = "systems@mlcommons.org" } diff --git a/script/build-dockerfile/customize.py b/script/build-dockerfile/customize.py index 50a4c987c..8a01ef753 100644 --- a/script/build-dockerfile/customize.py +++ b/script/build-dockerfile/customize.py @@ -4,6 +4,7 @@ import json import re import shutil +from utils import * def preprocess(i): @@ -255,8 +256,7 @@ def preprocess(i): 'MLC_DOCKER_USE_DEFAULT_USER', '') == '': env['MLC_DOCKER_USE_DEFAULT_USER'] = 'yes' - if docker_user and str(env.get('MLC_DOCKER_USE_DEFAULT_USER', '')).lower() not in [ - "yes", "1", "true"]: + if docker_user and not is_true(env.get('MLC_DOCKER_USE_DEFAULT_USER', '')): f.write('RUN groupadd -g $GID -o ' + docker_group + EOL) @@ -283,16 +283,19 @@ def preprocess(i): dockerfile_env_input_string = dockerfile_env_input_string + " --env." + \ docker_env_key + "=" + str(dockerfile_env[docker_env_key]) - workdir = get_value(env, config, 'WORKDIR', 'MLC_DOCKER_WORKDIR') - if workdir and (f"""/home/{docker_user}""" not in workdir or str(env.get('MLC_DOCKER_USE_DEFAULT_USER', '')).lower() not in [ - "yes", "1", "true"]): + workdir = env.get('WORKDIR', '') + if workdir == '': + workdir = f"""/home/{docker_user}""" + + if f"""/home/{docker_user}""" not in workdir or not is_true( + env.get('MLC_DOCKER_USE_DEFAULT_USER', '')): f.write('WORKDIR ' + workdir + EOL) f.write(EOL + '# Install python packages' + EOL) python = get_value(env, config, 'PYTHON', 'MLC_DOCKERFILE_PYTHON') docker_use_virtual_python = env.get('MLC_DOCKER_USE_VIRTUAL_PYTHON', "yes") - if str(docker_use_virtual_python).lower() not in ["no", "0", "false"]: + if not is_false(docker_use_virtual_python): f.write('RUN {} -m venv $HOME/venv/mlc'.format(python) + " " + EOL) f.write('ENV PATH="$HOME/venv/mlc/bin:$PATH"' + EOL) # f.write('RUN . /opt/venv/mlc/bin/activate' + EOL) @@ -342,8 +345,7 @@ def preprocess(i): for y in x.split(','): f.write('RUN ' + y + EOL) - if str(env.get('MLC_DOCKER_SKIP_MLC_SYS_UPGRADE', False) - ).lower() not in ["true", "1", "yes"]: + if not is_true(env.get('MLC_DOCKER_SKIP_MLC_SYS_UPGRADE', False)): f.write(EOL + '# Install all system dependencies' + EOL) f.write('RUN mlc run script --tags=get,sys-utils-mlc --quiet' + EOL) @@ -368,14 +370,12 @@ def preprocess(i): env['MLC_DOCKER_RUN_CMD'] += "mlc version" skip_extra = True else: - if str(env.get('MLC_DOCKER_NOT_PULL_UPDATE', 'False') - ).lower() not in ["yes", "1", "true"]: + if not is_true(env.get('MLC_DOCKER_NOT_PULL_UPDATE', 'False')): env['MLC_DOCKER_RUN_CMD'] += "mlc pull repo && " env['MLC_DOCKER_RUN_CMD'] += "mlc run script --tags=" + \ env['MLC_DOCKER_RUN_SCRIPT_TAGS'] + ' --quiet' else: - if str(env.get('MLC_DOCKER_NOT_PULL_UPDATE', 'False') - ).lower() not in ["yes", "1", "true"]: + if not is_true(env.get('MLC_DOCKER_NOT_PULL_UPDATE', 'False')): env['MLC_DOCKER_RUN_CMD'] = "mlc pull repo && " + \ env['MLC_DOCKER_RUN_CMD'] @@ -394,8 +394,8 @@ def preprocess(i): if run_cmd_extra != '': x += ' ' + run_cmd_extra - if env.get('MLC_DOCKER_RUN_SCRIPT_TAGS', '') != '' and str(env.get( - 'MLC_DOCKER_ADD_DEPENDENT_SCRIPTS_RUN_COMMANDS', '')).lower() in ["yes", "1", "true"]: + if env.get('MLC_DOCKER_RUN_SCRIPT_TAGS', '') != '' and is_true(env.get( + 'MLC_DOCKER_ADD_DEPENDENT_SCRIPTS_RUN_COMMANDS', '')): mlc_input = {'action': 'run', 'automation': 'script', 'tags': f"""{env['MLC_DOCKER_RUN_SCRIPT_TAGS']}""", @@ -417,8 +417,8 @@ def preprocess(i): f.write(x + EOL) # fake_run to install the dependent scripts and caching them - if not "run" in env['MLC_DOCKER_RUN_CMD'] and str( - env.get('MLC_REAL_RUN', False)).lower() in ["false", "0", "no"]: + if not "run" in env['MLC_DOCKER_RUN_CMD'] and is_false( + env.get('MLC_REAL_RUN', False)): fake_run = dockerfile_env_input_string x = 'RUN ' + env['MLC_DOCKER_RUN_CMD'] + fake_run + run_cmd_extra diff --git a/script/build-dockerfile/dockerinfo.json b/script/build-dockerfile/dockerinfo.json index cfb739e77..2b222c01c 100644 --- a/script/build-dockerfile/dockerinfo.json +++ b/script/build-dockerfile/dockerinfo.json @@ -22,11 +22,9 @@ "GID": "", "GROUP": "mlc", "SHELL": "[\"/bin/bash\", \"-c\"]", - "WORKDIR": "/home/mlcuser", "distros": { "ubuntu": { "USER": "ubuntu", - "WORKDIR": "/home/ubuntu", "package-manager-update-cmd": "apt-get update -y", "package-manager-get-cmd": "apt-get install -y", "packages": [ diff --git a/setup.py b/setup.py index c8d014c5c..5a18472c6 100644 --- a/setup.py +++ b/setup.py @@ -69,11 +69,11 @@ def run(self): print("Running custom post-install command...") commit_hash = get_commit_hash() import mlc - branch = os.environ.get('MLC_REPO_BRANCH', 'mlc') + branch = os.environ.get('MLC_REPO_BRANCH', 'dev') res = mlc.access({'action': 'pull', - 'automation': 'repo', - 'url': 'mlcommons@mlperf-automations', + 'target': 'repo', + 'repo': 'mlcommons@mlperf-automations', 'branch': branch, 'checkout': commit_hash })