Set 5 min timeout for LLM calls (#358)
This commit is contained in:
@@ -64,6 +64,7 @@ class Settings(BaseSettings):
|
|||||||
# ACTIVE LLM PROVIDER
|
# ACTIVE LLM PROVIDER
|
||||||
LLM_KEY: str = "OPENAI_GPT4O"
|
LLM_KEY: str = "OPENAI_GPT4O"
|
||||||
# COMMON
|
# COMMON
|
||||||
|
LLM_CONFIG_TIMEOUT: int = 300
|
||||||
LLM_CONFIG_MAX_TOKENS: int = 4096
|
LLM_CONFIG_MAX_TOKENS: int = 4096
|
||||||
LLM_CONFIG_TEMPERATURE: float = 0
|
LLM_CONFIG_TEMPERATURE: float = 0
|
||||||
# LLM PROVIDER SPECIFIC
|
# LLM PROVIDER SPECIFIC
|
||||||
|
|||||||
@@ -190,8 +190,10 @@ class LLMAPIHandlerFactory:
|
|||||||
response = await litellm.acompletion(
|
response = await litellm.acompletion(
|
||||||
model=llm_config.model_name,
|
model=llm_config.model_name,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
|
timeout=SettingsManager.get_settings().LLM_CONFIG_TIMEOUT,
|
||||||
**active_parameters,
|
**active_parameters,
|
||||||
)
|
)
|
||||||
|
LOG.info("LLM API call successful", llm_key=llm_key, model=llm_config.model_name)
|
||||||
except openai.OpenAIError as e:
|
except openai.OpenAIError as e:
|
||||||
raise LLMProviderError(llm_key) from e
|
raise LLMProviderError(llm_key) from e
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
|||||||
Reference in New Issue
Block a user