# import pytest # # from text_generation import ( # InferenceAPIClient, # InferenceAPIAsyncClient, # Client, # AsyncClient, # ) # from text_generation.errors import NotSupportedError, NotFoundError # from text_generation.inference_api import check_model_support, deployed_models # # # def test_check_model_support(flan_t5_xxl, unsupported_model, fake_model): # assert check_model_support(flan_t5_xxl) # assert not check_model_support(unsupported_model) # # with pytest.raises(NotFoundError): # check_model_support(fake_model) # # # def test_deployed_models(): # deployed_models() # # # def test_client(flan_t5_xxl): # client = InferenceAPIClient(flan_t5_xxl) # assert isinstance(client, Client) # # # def test_client_unsupported_model(unsupported_model): # with pytest.raises(NotSupportedError): # InferenceAPIClient(unsupported_model) # # # def test_async_client(flan_t5_xxl): # client = InferenceAPIAsyncClient(flan_t5_xxl) # assert isinstance(client, AsyncClient) # # # def test_async_client_unsupported_model(unsupported_model): # with pytest.raises(NotSupportedError): # InferenceAPIAsyncClient(unsupported_model)