@@ -68,7 +68,7 @@ def disable_embeddings():
6868 ModelProvider .ANTHROPIC : "claude-haiku-4-5" ,
6969 ModelProvider .AZURE : "gpt-4.1" , # TODO: Not validated yet.
7070 ModelProvider .GOOGLE : "gemini-2.5-flash" , # TODO: Not validated yet.
71- ModelProvider .HUGGINGFACE_API : "HuggingFaceH4/zephyr-7b-alpha" , # TODO: Not validated yet.
71+ ModelProvider .HUGGINGFACE_SERVERLESS : "HuggingFaceH4/zephyr-7b-alpha" , # TODO: Not validated yet.
7272 ModelProvider .OLLAMA : "gemma3:1b" ,
7373 ModelProvider .OPENAI : "gpt-4o-mini" ,
7474 ModelProvider .OPENROUTER : "gryphe/mythomax-l2-13b" ,
@@ -132,7 +132,7 @@ def read_llm_options(
132132 raise ValueError (
133133 "LLM API key not defined. Use either CLI/API parameter or GOOGLE_API_KEY environment variable."
134134 )
135- elif provider is ModelProvider .HUGGINGFACE_API :
135+ elif provider is ModelProvider .HUGGINGFACE_SERVERLESS :
136136 llm_api_key = llm_api_key or os .getenv ("HF_TOKEN" )
137137 if not llm_api_key :
138138 raise ValueError (
@@ -252,7 +252,7 @@ def configure_llm(info: ModelInfo, debug: bool = False) -> "LLM":
252252 base_url = info .endpoint ,
253253 api_key = info .api_key ,
254254 )
255- elif info .provider is ModelProvider .HUGGINGFACE_API :
255+ elif info .provider is ModelProvider .HUGGINGFACE_SERVERLESS :
256256 from llama_index .llms .huggingface_api import HuggingFaceInferenceAPI
257257
258258 llm = HuggingFaceInferenceAPI (
0 commit comments