.. |
__init__.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
bloom_modeling.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
clip.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
flash_cohere_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_dbrx_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_deepseek_v2_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_deepseek_v3_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_gemma2_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_gemma_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_gpt2_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_gptj_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_llama4_modeling.py
|
Enable Llama4 for Gaudi backend (#3223)
|
2025-05-15 14:35:37 +02:00 |
flash_llama_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_llava_next.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
flash_mistral_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_mixtral_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_mllama.py
|
Warmup gaudi backend (#3172)
|
2025-04-24 09:57:08 +02:00 |
flash_neox_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_pali_gemma_modeling.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
flash_phi_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_phi_moe_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_qwen2_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_rw_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_santacoder_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
flash_starcoder2_modeling.py
|
Deepseek R1 for Gaudi backend (#3211)
|
2025-05-19 16:36:39 +02:00 |
idefics2.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
idefics3.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
idefics_config.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
idefics_image_processing.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
idefics_modeling.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
idefics_perceiver.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
idefics_processing.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
idefics_vision.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
llava_next.py
|
Gaudi: Fix llava-next and mllama crash issue (#3127)
|
2025-03-25 15:08:15 +01:00 |
mamba_modeling.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
mllama.py
|
Gaudi: Sync TGI with the latest changes from the TGI-Gaudi fork (#3117)
|
2025-03-18 09:45:52 +01:00 |
qwen2_5_vl.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
qwen2_vl.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |
siglip.py
|
Add Gaudi Backend (#3055)
|
2025-02-28 12:14:58 +01:00 |
vlm.py
|
Gaudi: clean cuda/rocm code in hpu backend, enable flat_hpu (#3113)
|
2025-04-14 15:58:13 +02:00 |