From 46ccffd24689694909c296aab865c9ff573fddf7 Mon Sep 17 00:00:00 2001 From: Nicolas Patry Date: Wed, 9 Oct 2024 12:08:30 +0200 Subject: [PATCH] Flash llama on intel CPU ? --- .github/workflows/build.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/build.yaml b/.github/workflows/build.yaml index a48233dd..2e3ce8f1 100644 --- a/.github/workflows/build.yaml +++ b/.github/workflows/build.yaml @@ -77,7 +77,7 @@ jobs: export docker_volume="/mnt/cache" export runs_on="aws-highmemory-32-plus-priv" export platform="cpu" - export extra_pytest="-k test_flash_gemma_gptq_load" + export extra_pytest="-k test_flash_llama_load" ;; esac echo $dockerfile