From f80fe87b3f12f29ae8a196e0ba3fa0f5a4d881fc Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Thu, 7 Nov 2024 16:23:19 +0000 Subject: [PATCH 1/3] Update and rename test-mlperf-inference-intel to test-mlperf-inference-intel.yml --- ...rf-inference-intel => test-mlperf-inference-intel.yml} | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) rename .github/workflows/{test-mlperf-inference-intel => test-mlperf-inference-intel.yml} (85%) diff --git a/.github/workflows/test-mlperf-inference-intel b/.github/workflows/test-mlperf-inference-intel.yml similarity index 85% rename from .github/workflows/test-mlperf-inference-intel rename to .github/workflows/test-mlperf-inference-intel.yml index b52a7ff708..2a8e452f33 100644 --- a/.github/workflows/test-mlperf-inference-intel +++ b/.github/workflows/test-mlperf-inference-intel.yml @@ -2,7 +2,7 @@ name: MLPerf Inference Intel implementations on: schedule: - - cron: "54 14 * * *" #to be adjusted + - cron: "29 16 * * *" #to be adjusted jobs: build_nvidia: @@ -16,9 +16,9 @@ jobs: steps: - name: Test MLPerf Inference Intel ${{ matrix.model }} run: | - if [ -f "gh_action/bin/deactivate" ]; then source gh_action/bin/deactivate; fi - python3 -m venv gh_action - source gh_action/bin/activate + if [ -f "gh_action_conda/bin/deactivate" ]; then source gh_action_conda/bin/deactivate; fi + python3 -m venv gh_action_conda + source gh_action_conda/bin/activate export CM_REPOS=$HOME/GH_CM pip install --upgrade cm4mlops cm run script --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r4.1-dev --preprocess_submission=yes --execution_mode=valid --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=IntelSPR.24c --implementation=intel --backend=pytorch --category=datacenter --division=open --scenario=Offline --docker_dt=yes --docker_it=no --docker_cm_repo=gateoverflow@cm4mlops --adr.compiler.tags=gcc --device=cpu --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean --docker --quiet From 5200fcb4857678e03a510735fef753bbdb7b7e2b Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Thu, 7 Nov 2024 21:56:22 +0530 Subject: [PATCH 2/3] Support sample_ids_path in coco2014 accuracy script --- script/process-mlperf-accuracy/customize.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/script/process-mlperf-accuracy/customize.py b/script/process-mlperf-accuracy/customize.py index 4a6ba1634c..bb5d7a2865 100644 --- a/script/process-mlperf-accuracy/customize.py +++ b/script/process-mlperf-accuracy/customize.py @@ -103,6 +103,9 @@ def preprocess(i): else: extra_options += f""" --compliance-images-path '{os.path.join(result_dir, "images")}' """ + if env.get('CM_COCO2014_SAMPLE_ID_PATH','') != '': + extra_options += f" --ids-path '{env['CM_COCO2014_SAMPLE_ID_PATH']}' " + if env.get('CM_SDXL_ACCURACY_RUN_DEVICE', '') != '': extra_options += f" --device '{env['CM_SDXL_ACCURACY_RUN_DEVICE']}' " From 0600c69e759287fc6c6fe9e0a4ff2f1111538ad5 Mon Sep 17 00:00:00 2001 From: Arjun Suresh Date: Thu, 7 Nov 2024 22:02:03 +0530 Subject: [PATCH 3/3] Added rocm device for AMD mlperf inference --- script/app-mlperf-inference-amd/_cm.yaml | 10 ++++++++++ 1 file changed, 10 insertions(+) diff --git a/script/app-mlperf-inference-amd/_cm.yaml b/script/app-mlperf-inference-amd/_cm.yaml index 089a1be1eb..3882380372 100644 --- a/script/app-mlperf-inference-amd/_cm.yaml +++ b/script/app-mlperf-inference-amd/_cm.yaml @@ -146,6 +146,12 @@ variations: CM_MLPERF_DEVICE: gpu CM_MLPERF_DEVICE_LIB_NAMESPEC: cudart + rocm: + group: device + env: + CM_MLPERF_DEVICE: rocm + CM_MLPERF_DEVICE_LIB_NAMESPEC: rocm + openshift: group: backend default: true @@ -161,6 +167,10 @@ variations: deps: - tags: get,generic-python-lib,_torch_cuda + pytorch,rocm: + deps: + - tags: get,generic-python-lib,_torch,_rocm + pytorch,cpu: deps: - tags: get,generic-python-lib,_torch