From 2c837c671d6c3a6627c01952f5d04547a273922e Mon Sep 17 00:00:00 2001 From: baptiste Date: Fri, 4 Apr 2025 07:45:13 +0000 Subject: [PATCH] remove debug comments --- .github/workflows/build.yaml | 4 ++-- .../server/integration-tests/conftest.py | 19 ------------------- .../server/integration-tests/requirements.txt | 13 ------------- .../server/integration-tests/test_model.py | 8 -------- 4 files changed, 2 insertions(+), 42 deletions(-) diff --git a/.github/workflows/build.yaml b/.github/workflows/build.yaml index 1fbe6698..99f29d7e 100644 --- a/.github/workflows/build.yaml +++ b/.github/workflows/build.yaml @@ -129,9 +129,9 @@ jobs: export label_extension="-gaudi" export docker_volume="/mnt/cache" export docker_devices="" - export runs_on="aws-dl1-24xlarge" + export runs_on="ubuntu-latest" export platform="" - export extra_pytest="--gaudi" + export extra_pytest="" export target="" esac echo $dockerfile diff --git a/backends/gaudi/server/integration-tests/conftest.py b/backends/gaudi/server/integration-tests/conftest.py index aa50f833..c7daf70e 100644 --- a/backends/gaudi/server/integration-tests/conftest.py +++ b/backends/gaudi/server/integration-tests/conftest.py @@ -54,25 +54,6 @@ logger.add( level="INFO", ) -# def cleanup_handler(signum, frame): -# logger.info("\nCleaning up containers due to shutdown, please wait...") -# try: -# client = docker.from_env() -# containers = client.containers.list(filters={"name": "tgi-tests-"}) -# for container in containers: -# try: -# container.stop() -# container.remove() -# logger.info(f"Successfully cleaned up container {container.name}") -# except Exception as e: -# logger.error(f"Error cleaning up container {container.name}: {str(e)}") -# except Exception as e: -# logger.error(f"Error during cleanup: {str(e)}") -# sys.exit(1) - -# signal.signal(signal.SIGINT, cleanup_handler) -# signal.signal(signal.SIGTERM, cleanup_handler) - def stream_container_logs(container, test_name): """Stream container logs in a separate thread.""" diff --git a/backends/gaudi/server/integration-tests/requirements.txt b/backends/gaudi/server/integration-tests/requirements.txt index 320bfb7f..b67d2d8c 100644 --- a/backends/gaudi/server/integration-tests/requirements.txt +++ b/backends/gaudi/server/integration-tests/requirements.txt @@ -1,16 +1,3 @@ -# Copyright 2023 The HuggingFace Team. All rights reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. pytest >= 8.3.5 pytest-asyncio >= 0.26.0 docker >= 7.1.0 diff --git a/backends/gaudi/server/integration-tests/test_model.py b/backends/gaudi/server/integration-tests/test_model.py index 7185a56a..7212faea 100644 --- a/backends/gaudi/server/integration-tests/test_model.py +++ b/backends/gaudi/server/integration-tests/test_model.py @@ -4,14 +4,6 @@ from text_generation import AsyncClient import pytest from Levenshtein import distance as levenshtein_distance -# Model that are not working but should... :( -# - google/gemma-2-2b-it -# - google/flan-t5-large -# - codellama/CodeLlama-13b-hf -# - ibm-granite/granite-3.0-8b-instruct -# - microsoft/Phi-3.5-MoE-instruct -# - microsoft/Phi-3-mini-4k-instruct - # The config in args is not optimized for speed but only check that inference is working for the different models architectures TEST_CONFIGS = { "meta-llama/Llama-3.1-8B-Instruct-shared": {