fallback gemini to gpt + use latest gemini flash 2.5 release (#2588)
This commit is contained in:
@@ -833,6 +833,21 @@ if settings.ENABLE_VERTEX_AI and settings.VERTEX_CREDENTIALS:
|
|||||||
)
|
)
|
||||||
LLMConfigRegistry.register_config(
|
LLMConfigRegistry.register_config(
|
||||||
"VERTEX_GEMINI_2.5_FLASH_PREVIEW",
|
"VERTEX_GEMINI_2.5_FLASH_PREVIEW",
|
||||||
|
LLMConfig(
|
||||||
|
"vertex_ai/gemini-2.5-flash-preview-05-20",
|
||||||
|
["VERTEX_CREDENTIALS"],
|
||||||
|
supports_vision=True,
|
||||||
|
add_assistant_prefix=False,
|
||||||
|
max_completion_tokens=65535,
|
||||||
|
litellm_params=LiteLLMParams(
|
||||||
|
vertex_credentials=settings.VERTEX_CREDENTIALS,
|
||||||
|
api_base=f"{api_base}/gemini-2.5-flash-preview-05-20" if api_base else None,
|
||||||
|
vertex_location=settings.VERTEX_LOCATION,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
LLMConfigRegistry.register_config(
|
||||||
|
"VERTEX_GEMINI_2.5_FLASH_PREVIEW_04_17",
|
||||||
LLMConfig(
|
LLMConfig(
|
||||||
"vertex_ai/gemini-2.5-flash-preview-04-17",
|
"vertex_ai/gemini-2.5-flash-preview-04-17",
|
||||||
["VERTEX_CREDENTIALS"],
|
["VERTEX_CREDENTIALS"],
|
||||||
|
|||||||
Reference in New Issue
Block a user