From 1bb1a344d79ce1b2c3d648c31b3683426ab98d72 Mon Sep 17 00:00:00 2001 From: Felix Marty <9808326+fxmarty@users.noreply.github.com> Date: Fri, 21 Jun 2024 16:11:38 +0000 Subject: [PATCH] retry --- integration-tests/clean_cache_and_download.py | 7 +------ 1 file changed, 1 insertion(+), 6 deletions(-) diff --git a/integration-tests/clean_cache_and_download.py b/integration-tests/clean_cache_and_download.py index d6d79aa3..d1011240 100644 --- a/integration-tests/clean_cache_and_download.py +++ b/integration-tests/clean_cache_and_download.py @@ -40,15 +40,14 @@ def cleanup_cache(token: str): size_per_model = {} extension_per_model = {} for model_id, revision in REQUIRED_MODELS.items(): + print(f"Crawling {model_id}...") model_size = 0 - print(f"call huggingface_hub.list_repo_files for {model_id}") all_files = huggingface_hub.list_repo_files( model_id, repo_type="model", revision=revision, token=token, ) - print(f"end huggingface_hub.list_repo_files") extension = None if any(".safetensors" in filename for filename in all_files): @@ -60,9 +59,6 @@ def cleanup_cache(token: str): extension_per_model[model_id] = extension - print( - f"call huggingface_hub.hf_hub_url & huggingface_hub.get_hf_file_metadata for {model_id}" - ) for filename in all_files: if filename.endswith(extension): file_url = huggingface_hub.hf_hub_url( @@ -72,7 +68,6 @@ def cleanup_cache(token: str): file_url, token=token ) model_size += file_metadata.size * 1e-9 # in GB - print(f"end two calls") size_per_model[model_id] = model_size