From 398785ce2e27344305fc41e4debfd4547cb102c8 Mon Sep 17 00:00:00 2001 From: API Engineering Date: Tue, 28 Apr 2026 10:33:13 +0000 Subject: [PATCH] [bot] Updated client based on openapi-9485b4a/clientgen --- DO_OPENAPI_COMMIT_SHA.txt | 2 +- src/pydo/aio/operations/_operations.py | 24147 +++++++++++++++++++--- src/pydo/operations/_operations.py | 24460 ++++++++++++++++++++--- 3 files changed, 43048 insertions(+), 5561 deletions(-) diff --git a/DO_OPENAPI_COMMIT_SHA.txt b/DO_OPENAPI_COMMIT_SHA.txt index b8c97a26..7e399b9c 100644 --- a/DO_OPENAPI_COMMIT_SHA.txt +++ b/DO_OPENAPI_COMMIT_SHA.txt @@ -1 +1 @@ -49946b8 +9485b4a diff --git a/src/pydo/aio/operations/_operations.py b/src/pydo/aio/operations/_operations.py index 817722cb..36883aea 100644 --- a/src/pydo/aio/operations/_operations.py +++ b/src/pydo/aio/operations/_operations.py @@ -269,6 +269,8 @@ build_genai_create_knowledge_base_data_source_request, build_genai_create_knowledge_base_request, build_genai_create_model_api_key_request, + build_genai_create_model_eval_dataset_upload_presigned_urls_request, + build_genai_create_model_evaluation_run_request, build_genai_create_model_router_request, build_genai_create_oauth2_dropbox_tokens_request, build_genai_create_openai_api_key_request, @@ -277,6 +279,7 @@ build_genai_delete_agent_api_key_request, build_genai_delete_agent_request, build_genai_delete_anthropic_api_key_request, + build_genai_delete_custom_model_request, build_genai_delete_knowledge_base_data_source_request, build_genai_delete_knowledge_base_request, build_genai_delete_model_api_key_request, @@ -292,6 +295,7 @@ build_genai_get_agent_request, build_genai_get_agent_usage_request, build_genai_get_anthropic_api_key_request, + build_genai_get_evaluation_dataset_download_url_request, build_genai_get_evaluation_run_prompt_results_request, build_genai_get_evaluation_run_request, build_genai_get_evaluation_run_results_request, @@ -299,11 +303,15 @@ build_genai_get_indexing_job_details_signed_url_request, build_genai_get_indexing_job_request, build_genai_get_knowledge_base_request, + build_genai_get_model_catalog_card_request, + build_genai_get_model_evaluation_run_request, + build_genai_get_model_evaluation_run_results_download_url_request, build_genai_get_model_router_request, build_genai_get_oauth2_url_request, build_genai_get_openai_api_key_request, build_genai_get_scheduled_indexing_request, build_genai_get_workspace_request, + build_genai_import_custom_model_request, build_genai_list_agent_api_keys_request, build_genai_list_agent_versions_request, build_genai_list_agents_by_anthropic_key_request, @@ -311,6 +319,7 @@ build_genai_list_agents_by_workspace_request, build_genai_list_agents_request, build_genai_list_anthropic_api_keys_request, + build_genai_list_custom_models_request, build_genai_list_datacenter_regions_request, build_genai_list_evaluation_metrics_request, build_genai_list_evaluation_runs_by_test_case_request, @@ -322,6 +331,9 @@ build_genai_list_knowledge_base_data_sources_request, build_genai_list_knowledge_bases_request, build_genai_list_model_api_keys_request, + build_genai_list_model_catalog_request, + build_genai_list_model_evaluation_metrics_request, + build_genai_list_model_evaluation_runs_request, build_genai_list_model_router_presets_request, build_genai_list_model_router_task_presets_request, build_genai_list_model_routers_request, @@ -339,6 +351,7 @@ build_genai_update_agents_workspace_request, build_genai_update_anthropic_api_key_request, build_genai_update_attached_agent_request, + build_genai_update_custom_model_metadata_request, build_genai_update_evaluation_test_case_request, build_genai_update_knowledge_base_data_source_request, build_genai_update_knowledge_base_request, @@ -200523,6 +200536,24 @@ async def list_agents( "max_tokens": 0, # Optional. Specifies the maximum number of tokens the model can process in a single input or output, set as a number between 1 and 512. This determines the length of each response. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -200534,8 +200565,27 @@ async def list_agents( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -200548,14 +200598,59 @@ async def list_agents( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -200573,11 +200668,52 @@ async def list_agents( "patch": 0 # Optional. Patch version number. } }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "project_id": "str", # Optional. The DigitalOcean project ID associated with the agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -200762,8 +200898,27 @@ async def list_agents( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -200777,16 +200932,64 @@ async def list_agents( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -200830,6 +201033,8 @@ async def list_agents( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Defines the cumulative probability threshold for word selection, specified as a number between 0 and 1. Higher values allow for more diverse outputs, while lower values ensure @@ -200990,18 +201195,39 @@ async def create_agent( "str" # Optional. Ids of the knowledge base(s) to attach to the agent. ], + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI API key ID to use with OpenAI models. "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. + "reasoning_effort": "str", # Optional. "region": "str", # Optional. The DigitalOcean region to deploy your agent in. + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. Agent tag to organize related resources. ], + "thinking_token_budget": 0, # Optional. "workspace_uuid": "str" # Optional. Identifier for the workspace. } @@ -201274,6 +201500,24 @@ async def create_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -201282,8 +201526,27 @@ async def create_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -201295,14 +201558,56 @@ async def create_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -201337,8 +201642,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -201352,16 +201676,64 @@ async def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -201390,6 +201762,44 @@ async def create_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -201409,8 +201819,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -201424,16 +201853,64 @@ async def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -201465,6 +201942,8 @@ async def create_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -201642,8 +202121,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -201656,14 +202154,59 @@ async def create_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -201703,6 +202246,8 @@ async def create_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -201742,6 +202287,13 @@ async def create_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -201772,6 +202324,15 @@ async def create_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -201785,7 +202346,8 @@ async def create_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -202141,6 +202703,24 @@ async def create_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -202149,8 +202729,27 @@ async def create_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -202162,14 +202761,56 @@ async def create_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -202204,8 +202845,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -202219,16 +202879,64 @@ async def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -202257,6 +202965,44 @@ async def create_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -202276,8 +203022,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -202291,16 +203056,64 @@ async def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -202332,6 +203145,8 @@ async def create_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -202509,8 +203324,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -202523,14 +203357,59 @@ async def create_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -202570,6 +203449,8 @@ async def create_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -202609,6 +203490,13 @@ async def create_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -202639,6 +203527,15 @@ async def create_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -202652,7 +203549,8 @@ async def create_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -202746,18 +203644,39 @@ async def create_agent( "str" # Optional. Ids of the knowledge base(s) to attach to the agent. ], + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI API key ID to use with OpenAI models. "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. + "reasoning_effort": "str", # Optional. "region": "str", # Optional. The DigitalOcean region to deploy your agent in. + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. Agent tag to organize related resources. ], + "thinking_token_budget": 0, # Optional. "workspace_uuid": "str" # Optional. Identifier for the workspace. } @@ -203030,6 +203949,24 @@ async def create_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -203038,8 +203975,27 @@ async def create_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -203051,14 +204007,56 @@ async def create_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -203093,8 +204091,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -203108,16 +204125,64 @@ async def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -203146,6 +204211,44 @@ async def create_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -203165,8 +204268,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -203180,16 +204302,64 @@ async def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -203221,6 +204391,8 @@ async def create_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -203398,8 +204570,27 @@ async def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -203412,14 +204603,59 @@ async def create_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -203459,6 +204695,8 @@ async def create_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -203498,6 +204736,13 @@ async def create_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -203528,6 +204773,15 @@ async def create_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -203541,7 +204795,8 @@ async def create_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -204930,6 +206185,24 @@ async def attach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -204938,8 +206211,27 @@ async def attach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -204951,14 +206243,56 @@ async def attach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -204993,8 +206327,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -205008,16 +206361,64 @@ async def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -205046,6 +206447,44 @@ async def attach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -205065,8 +206504,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -205080,16 +206538,64 @@ async def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -205121,6 +206627,8 @@ async def attach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -205298,8 +206806,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -205312,14 +206839,59 @@ async def attach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -205359,6 +206931,8 @@ async def attach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -205398,6 +206972,13 @@ async def attach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -205428,6 +207009,15 @@ async def attach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -205441,7 +207031,8 @@ async def attach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -205800,6 +207391,24 @@ async def attach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -205808,8 +207417,27 @@ async def attach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -205821,14 +207449,56 @@ async def attach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -205863,8 +207533,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -205878,16 +207567,64 @@ async def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -205916,6 +207653,44 @@ async def attach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -205935,8 +207710,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -205950,16 +207744,64 @@ async def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -205991,6 +207833,8 @@ async def attach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -206168,8 +208012,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -206182,14 +208045,59 @@ async def attach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -206229,6 +208137,8 @@ async def attach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -206268,6 +208178,13 @@ async def attach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -206298,6 +208215,15 @@ async def attach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -206311,7 +208237,8 @@ async def attach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -206680,6 +208607,24 @@ async def attach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -206688,8 +208633,27 @@ async def attach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -206701,14 +208665,56 @@ async def attach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -206743,8 +208749,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -206758,16 +208783,64 @@ async def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -206796,6 +208869,44 @@ async def attach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -206815,8 +208926,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -206830,16 +208960,64 @@ async def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -206871,6 +209049,8 @@ async def attach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -207048,8 +209228,27 @@ async def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -207062,14 +209261,59 @@ async def attach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -207109,6 +209353,8 @@ async def attach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -207148,6 +209394,13 @@ async def attach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -207178,6 +209431,15 @@ async def attach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -207191,7 +209453,8 @@ async def attach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -207663,6 +209926,24 @@ async def update_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -207671,8 +209952,27 @@ async def update_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -207684,14 +209984,56 @@ async def update_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -207726,8 +210068,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -207741,16 +210102,64 @@ async def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -207779,6 +210188,44 @@ async def update_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -207798,8 +210245,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -207813,16 +210279,64 @@ async def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -207854,6 +210368,8 @@ async def update_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -208031,8 +210547,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -208045,14 +210580,59 @@ async def update_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -208092,6 +210672,8 @@ async def update_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -208131,6 +210713,13 @@ async def update_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -208161,6 +210750,15 @@ async def update_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -208174,7 +210772,8 @@ async def update_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -208536,6 +211135,24 @@ async def update_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -208544,8 +211161,27 @@ async def update_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -208557,14 +211193,56 @@ async def update_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -208599,8 +211277,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -208614,16 +211311,64 @@ async def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -208652,6 +211397,44 @@ async def update_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -208671,8 +211454,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -208686,16 +211488,64 @@ async def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -208727,6 +211577,8 @@ async def update_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -208904,8 +211756,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -208918,14 +211789,59 @@ async def update_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -208965,6 +211881,8 @@ async def update_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -209004,6 +211922,13 @@ async def update_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -209034,6 +211959,15 @@ async def update_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -209047,7 +211981,8 @@ async def update_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -209420,6 +212355,24 @@ async def update_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -209428,8 +212381,27 @@ async def update_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -209441,14 +212413,56 @@ async def update_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -209483,8 +212497,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -209498,16 +212531,64 @@ async def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -209536,6 +212617,44 @@ async def update_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -209555,8 +212674,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -209570,16 +212708,64 @@ async def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -209611,6 +212797,8 @@ async def update_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -209788,8 +212976,27 @@ async def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -209802,14 +213009,59 @@ async def update_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -209849,6 +213101,8 @@ async def update_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -209888,6 +213142,13 @@ async def update_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -209918,6 +213179,15 @@ async def update_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -209931,7 +213201,8 @@ async def update_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -210378,6 +213649,24 @@ async def detach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -210386,8 +213675,27 @@ async def detach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -210399,14 +213707,56 @@ async def detach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -210441,8 +213791,27 @@ async def detach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -210456,16 +213825,64 @@ async def detach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -210494,6 +213911,44 @@ async def detach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -210513,8 +213968,27 @@ async def detach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -210528,16 +214002,64 @@ async def detach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -210569,6 +214091,8 @@ async def detach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -210746,8 +214270,27 @@ async def detach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -210760,14 +214303,59 @@ async def detach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -210807,6 +214395,8 @@ async def detach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -210846,6 +214436,13 @@ async def detach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -210876,6 +214473,15 @@ async def detach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -210889,7 +214495,8 @@ async def detach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -211337,6 +214944,24 @@ async def attach_agent_guardrails( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -211345,8 +214970,27 @@ async def attach_agent_guardrails( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -211358,14 +215002,56 @@ async def attach_agent_guardrails( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -211400,8 +215086,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -211415,16 +215120,64 @@ async def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -211453,6 +215206,44 @@ async def attach_agent_guardrails( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -211472,8 +215263,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -211487,16 +215297,64 @@ async def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -211528,6 +215386,8 @@ async def attach_agent_guardrails( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -211705,8 +215565,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -211719,14 +215598,59 @@ async def attach_agent_guardrails( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -211766,6 +215690,8 @@ async def attach_agent_guardrails( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -211805,6 +215731,13 @@ async def attach_agent_guardrails( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -211835,6 +215768,15 @@ async def attach_agent_guardrails( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -211848,7 +215790,8 @@ async def attach_agent_guardrails( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -212207,6 +216150,24 @@ async def attach_agent_guardrails( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -212215,8 +216176,27 @@ async def attach_agent_guardrails( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -212228,14 +216208,56 @@ async def attach_agent_guardrails( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -212270,8 +216292,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -212285,16 +216326,64 @@ async def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -212323,6 +216412,44 @@ async def attach_agent_guardrails( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -212342,8 +216469,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -212357,16 +216503,64 @@ async def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -212398,6 +216592,8 @@ async def attach_agent_guardrails( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -212575,8 +216771,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -212589,14 +216804,59 @@ async def attach_agent_guardrails( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -212636,6 +216896,8 @@ async def attach_agent_guardrails( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -212675,6 +216937,13 @@ async def attach_agent_guardrails( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -212705,6 +216974,15 @@ async def attach_agent_guardrails( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -212718,7 +216996,8 @@ async def attach_agent_guardrails( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -213083,6 +217362,24 @@ async def attach_agent_guardrails( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -213091,8 +217388,27 @@ async def attach_agent_guardrails( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -213104,14 +217420,56 @@ async def attach_agent_guardrails( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -213146,8 +217504,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -213161,16 +217538,64 @@ async def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -213199,6 +217624,44 @@ async def attach_agent_guardrails( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -213218,8 +217681,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -213233,16 +217715,64 @@ async def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -213274,6 +217804,8 @@ async def attach_agent_guardrails( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -213451,8 +217983,27 @@ async def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -213465,14 +218016,59 @@ async def attach_agent_guardrails( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -213512,6 +218108,8 @@ async def attach_agent_guardrails( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -213551,6 +218149,13 @@ async def attach_agent_guardrails( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -213581,6 +218186,15 @@ async def attach_agent_guardrails( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -213594,7 +218208,8 @@ async def attach_agent_guardrails( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -214039,6 +218654,24 @@ async def detach_agent_guardrail( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -214047,8 +218680,27 @@ async def detach_agent_guardrail( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -214060,14 +218712,56 @@ async def detach_agent_guardrail( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -214102,8 +218796,27 @@ async def detach_agent_guardrail( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -214117,16 +218830,64 @@ async def detach_agent_guardrail( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -214155,6 +218916,44 @@ async def detach_agent_guardrail( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -214174,8 +218973,27 @@ async def detach_agent_guardrail( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -214189,16 +219007,64 @@ async def detach_agent_guardrail( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -214230,6 +219096,8 @@ async def detach_agent_guardrail( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -214407,8 +219275,27 @@ async def detach_agent_guardrail( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -214421,14 +219308,59 @@ async def detach_agent_guardrail( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -214468,6 +219400,8 @@ async def detach_agent_guardrail( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -214507,6 +219441,13 @@ async def detach_agent_guardrail( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -214537,6 +219478,15 @@ async def detach_agent_guardrail( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -214550,7 +219500,8 @@ async def detach_agent_guardrail( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -214975,6 +219926,24 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -214983,8 +219952,27 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -214996,14 +219984,56 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -215038,8 +220068,27 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -215053,16 +220102,64 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -215091,6 +220188,44 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -215110,8 +220245,27 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -215125,16 +220279,64 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -215166,6 +220368,8 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -215343,8 +220547,27 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -215357,14 +220580,59 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -215404,6 +220672,8 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -215443,6 +220713,13 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -215473,6 +220750,15 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -215486,7 +220772,8 @@ async def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -215914,6 +221201,24 @@ async def attach_knowledge_base( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -215922,8 +221227,27 @@ async def attach_knowledge_base( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -215935,14 +221259,56 @@ async def attach_knowledge_base( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -215977,8 +221343,27 @@ async def attach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -215992,16 +221377,64 @@ async def attach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -216030,6 +221463,44 @@ async def attach_knowledge_base( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -216049,8 +221520,27 @@ async def attach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -216064,16 +221554,64 @@ async def attach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -216105,6 +221643,8 @@ async def attach_knowledge_base( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -216282,8 +221822,27 @@ async def attach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -216296,14 +221855,59 @@ async def attach_knowledge_base( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -216343,6 +221947,8 @@ async def attach_knowledge_base( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -216382,6 +221988,13 @@ async def attach_knowledge_base( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -216412,6 +222025,15 @@ async def attach_knowledge_base( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -216425,7 +222047,8 @@ async def attach_knowledge_base( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -216854,6 +222477,24 @@ async def detach_knowledge_base( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -216862,8 +222503,27 @@ async def detach_knowledge_base( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -216875,14 +222535,56 @@ async def detach_knowledge_base( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -216917,8 +222619,27 @@ async def detach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -216932,16 +222653,64 @@ async def detach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -216970,6 +222739,44 @@ async def detach_knowledge_base( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -216989,8 +222796,27 @@ async def detach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -217004,16 +222830,64 @@ async def detach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -217045,6 +222919,8 @@ async def detach_knowledge_base( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -217222,8 +223098,27 @@ async def detach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -217236,14 +223131,59 @@ async def detach_knowledge_base( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -217283,6 +223223,8 @@ async def detach_knowledge_base( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -217322,6 +223264,13 @@ async def detach_knowledge_base( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -217352,6 +223301,15 @@ async def detach_knowledge_base( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -217365,7 +223323,8 @@ async def detach_knowledge_base( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -218439,6 +224398,24 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -218447,8 +224424,27 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -218460,14 +224456,56 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -218502,8 +224540,27 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -218517,16 +224574,64 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -218555,6 +224660,44 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -218574,8 +224717,27 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -218589,16 +224751,64 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -218630,6 +224840,8 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -218807,8 +225019,27 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -218821,14 +225052,59 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -218868,6 +225144,8 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -218907,6 +225185,13 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -218937,6 +225222,15 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -218950,7 +225244,8 @@ async def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -219126,6 +225421,8 @@ async def update_agent( ], "anthropic_key_uuid": "str", # Optional. Optional anthropic key uuid for use with anthropic models. + "clear_mcp_servers": bool, # Optional. When true, removes all MCP servers + from the agent. Use this instead of sending an empty mcp_servers array. "conversation_logs_enabled": bool, # Optional. Optional update of conversation logs enabled. "description": "str", # Optional. Agent description. @@ -219138,8 +225435,26 @@ async def update_agent( "max_tokens": 0, # Optional. Specifies the maximum number of tokens the model can process in a single input or output, set as a number between 1 and 512. This determines the length of each response. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. Optional Model Provider uuid for use with provider models. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI key uuid for use with @@ -219147,6 +225462,7 @@ async def update_agent( "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. "provide_citations": bool, # Optional. + "reasoning_effort": "str", # Optional. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The retrieval method is unknown * RETRIEVAL_METHOD_REWRITE: The retrieval method is rewrite * @@ -219156,12 +225472,14 @@ async def update_agent( "RETRIEVAL_METHOD_UNKNOWN", "RETRIEVAL_METHOD_REWRITE", "RETRIEVAL_METHOD_STEP_BACK", "RETRIEVAL_METHOD_SUB_QUERIES", and "RETRIEVAL_METHOD_NONE". + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. A set of abitrary tags to organize your agent. ], "temperature": 0.0, # Optional. Controls the model"u2019s creativity, specified as a number between 0 and 1. Lower values produce more predictable and conservative responses, while higher values encourage creativity and variation. + "thinking_token_budget": 0, # Optional. "top_p": 0.0, # Optional. Defines the cumulative probability threshold for word selection, specified as a number between 0 and 1. Higher values allow for more diverse outputs, while lower values ensure focused and coherent responses. @@ -219437,6 +225755,24 @@ async def update_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -219445,8 +225781,27 @@ async def update_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -219458,14 +225813,56 @@ async def update_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -219500,8 +225897,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -219515,16 +225931,64 @@ async def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -219553,6 +226017,44 @@ async def update_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -219572,8 +226074,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -219587,16 +226108,64 @@ async def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -219628,6 +226197,8 @@ async def update_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -219805,8 +226376,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -219819,14 +226409,59 @@ async def update_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -219866,6 +226501,8 @@ async def update_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -219905,6 +226542,13 @@ async def update_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -219935,6 +226579,15 @@ async def update_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -219948,7 +226601,8 @@ async def update_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -220307,6 +226961,24 @@ async def update_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -220315,8 +226987,27 @@ async def update_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -220328,14 +227019,56 @@ async def update_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -220370,8 +227103,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -220385,16 +227137,64 @@ async def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -220423,6 +227223,44 @@ async def update_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -220442,8 +227280,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -220457,16 +227314,64 @@ async def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -220498,6 +227403,8 @@ async def update_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -220675,8 +227582,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -220689,14 +227615,59 @@ async def update_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -220736,6 +227707,8 @@ async def update_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -220775,6 +227748,13 @@ async def update_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -220805,6 +227785,15 @@ async def update_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -220818,7 +227807,8 @@ async def update_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -220909,6 +227899,8 @@ async def update_agent( ], "anthropic_key_uuid": "str", # Optional. Optional anthropic key uuid for use with anthropic models. + "clear_mcp_servers": bool, # Optional. When true, removes all MCP servers + from the agent. Use this instead of sending an empty mcp_servers array. "conversation_logs_enabled": bool, # Optional. Optional update of conversation logs enabled. "description": "str", # Optional. Agent description. @@ -220921,8 +227913,26 @@ async def update_agent( "max_tokens": 0, # Optional. Specifies the maximum number of tokens the model can process in a single input or output, set as a number between 1 and 512. This determines the length of each response. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. Optional Model Provider uuid for use with provider models. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI key uuid for use with @@ -220930,6 +227940,7 @@ async def update_agent( "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. "provide_citations": bool, # Optional. + "reasoning_effort": "str", # Optional. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The retrieval method is unknown * RETRIEVAL_METHOD_REWRITE: The retrieval method is rewrite * @@ -220939,12 +227950,14 @@ async def update_agent( "RETRIEVAL_METHOD_UNKNOWN", "RETRIEVAL_METHOD_REWRITE", "RETRIEVAL_METHOD_STEP_BACK", "RETRIEVAL_METHOD_SUB_QUERIES", and "RETRIEVAL_METHOD_NONE". + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. A set of abitrary tags to organize your agent. ], "temperature": 0.0, # Optional. Controls the model"u2019s creativity, specified as a number between 0 and 1. Lower values produce more predictable and conservative responses, while higher values encourage creativity and variation. + "thinking_token_budget": 0, # Optional. "top_p": 0.0, # Optional. Defines the cumulative probability threshold for word selection, specified as a number between 0 and 1. Higher values allow for more diverse outputs, while lower values ensure focused and coherent responses. @@ -221220,6 +228233,24 @@ async def update_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -221228,8 +228259,27 @@ async def update_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -221241,14 +228291,56 @@ async def update_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -221283,8 +228375,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -221298,16 +228409,64 @@ async def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -221336,6 +228495,44 @@ async def update_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -221355,8 +228552,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -221370,16 +228586,64 @@ async def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -221411,6 +228675,8 @@ async def update_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -221588,8 +228854,27 @@ async def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -221602,14 +228887,59 @@ async def update_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -221649,6 +228979,8 @@ async def update_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -221688,6 +229020,13 @@ async def update_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -221718,6 +229057,15 @@ async def update_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -221731,7 +229079,8 @@ async def update_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -222171,6 +229520,24 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -222179,8 +229546,27 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -222192,14 +229578,56 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -222234,8 +229662,27 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -222249,16 +229696,64 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -222287,6 +229782,44 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -222306,8 +229839,27 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -222321,16 +229873,64 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -222362,6 +229962,8 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -222539,8 +230141,27 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -222553,14 +230174,59 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -222600,6 +230266,8 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -222639,6 +230307,13 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -222669,6 +230344,15 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -222682,7 +230366,8 @@ async def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -223128,6 +230813,24 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: stream. }, "max_tokens": 0, # Optional. Child agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -223139,8 +230842,27 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -223153,14 +230875,59 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -223198,8 +230965,27 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -223213,10 +230999,25 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -223224,6 +231025,42 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -223254,6 +231091,45 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -223274,8 +231150,27 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -223289,10 +231184,25 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -223300,6 +231210,42 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -223333,6 +231279,8 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. Child agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -223513,8 +231461,27 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -223528,16 +231495,64 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -223581,6 +231596,8 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Child agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -223623,6 +231640,13 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -223653,6 +231677,16 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -223666,8 +231700,10 @@ async def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -224122,6 +232158,24 @@ async def update_agent_deployment_visibility( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -224130,8 +232184,27 @@ async def update_agent_deployment_visibility( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -224143,14 +232216,56 @@ async def update_agent_deployment_visibility( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -224185,8 +232300,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -224200,16 +232334,64 @@ async def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -224238,6 +232420,44 @@ async def update_agent_deployment_visibility( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -224257,8 +232477,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -224272,16 +232511,64 @@ async def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -224313,6 +232600,8 @@ async def update_agent_deployment_visibility( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -224490,8 +232779,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -224504,14 +232812,59 @@ async def update_agent_deployment_visibility( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -224551,6 +232904,8 @@ async def update_agent_deployment_visibility( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -224590,6 +232945,13 @@ async def update_agent_deployment_visibility( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -224620,6 +232982,15 @@ async def update_agent_deployment_visibility( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -224633,7 +233004,8 @@ async def update_agent_deployment_visibility( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -224992,6 +233364,24 @@ async def update_agent_deployment_visibility( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -225000,8 +233390,27 @@ async def update_agent_deployment_visibility( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -225013,14 +233422,56 @@ async def update_agent_deployment_visibility( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -225055,8 +233506,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -225070,16 +233540,64 @@ async def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -225108,6 +233626,44 @@ async def update_agent_deployment_visibility( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -225127,8 +233683,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -225142,16 +233717,64 @@ async def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -225183,6 +233806,8 @@ async def update_agent_deployment_visibility( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -225360,8 +233985,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -225374,14 +234018,59 @@ async def update_agent_deployment_visibility( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -225421,6 +234110,8 @@ async def update_agent_deployment_visibility( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -225460,6 +234151,13 @@ async def update_agent_deployment_visibility( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -225490,6 +234188,15 @@ async def update_agent_deployment_visibility( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -225503,7 +234210,8 @@ async def update_agent_deployment_visibility( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -225868,6 +234576,24 @@ async def update_agent_deployment_visibility( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -225876,8 +234602,27 @@ async def update_agent_deployment_visibility( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -225889,14 +234634,56 @@ async def update_agent_deployment_visibility( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -225931,8 +234718,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -225946,16 +234752,64 @@ async def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -225984,6 +234838,44 @@ async def update_agent_deployment_visibility( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -226003,8 +234895,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -226018,16 +234929,64 @@ async def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -226059,6 +235018,8 @@ async def update_agent_deployment_visibility( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -226236,8 +235197,27 @@ async def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -226250,14 +235230,59 @@ async def update_agent_deployment_visibility( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -226297,6 +235322,8 @@ async def update_agent_deployment_visibility( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -226336,6 +235363,13 @@ async def update_agent_deployment_visibility( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -226366,6 +235400,15 @@ async def update_agent_deployment_visibility( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -226379,7 +235422,8 @@ async def update_agent_deployment_visibility( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -228405,6 +237449,24 @@ async def list_agents_by_anthropic_key( stream. }, "max_tokens": 0, # Optional. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -228416,8 +237478,27 @@ async def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -228430,14 +237511,59 @@ async def list_agents_by_anthropic_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -228475,8 +237601,27 @@ async def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -228490,10 +237635,25 @@ async def list_agents_by_anthropic_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -228501,6 +237661,42 @@ async def list_agents_by_anthropic_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -228531,6 +237727,45 @@ async def list_agents_by_anthropic_key( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -228551,8 +237786,27 @@ async def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -228566,10 +237820,25 @@ async def list_agents_by_anthropic_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -228577,6 +237846,42 @@ async def list_agents_by_anthropic_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -228610,6 +237915,8 @@ async def list_agents_by_anthropic_key( "project_id": "str", # Optional. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -228790,8 +238097,27 @@ async def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -228805,16 +238131,64 @@ async def list_agents_by_anthropic_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -228858,6 +238232,8 @@ async def list_agents_by_anthropic_key( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -228900,6 +238276,13 @@ async def list_agents_by_anthropic_key( 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -228930,6 +238313,16 @@ async def list_agents_by_anthropic_key( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -228943,8 +238336,10 @@ async def list_agents_by_anthropic_key( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -229102,81 +238497,28 @@ async def list_agents_by_anthropic_key( return cast(JSON, deserialized) # type: ignore - @overload - async def create_evaluation_dataset( - self, - body: Optional[JSON] = None, - *, - content_type: str = "application/json", - **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Create Evaluation Dataset. - - To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. - - :param body: Default value is None. - :type body: JSON - :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value - is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: - "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", and - "EVALUATION_DATASET_TYPE_NON_ADK". - "file_upload_dataset": { - "original_file_name": "str", # Optional. The original file name. - "size_in_bytes": "str", # Optional. The size of the file in bytes. - "stored_object_key": "str" # Optional. The object key the file was - stored as. - }, - "name": "str" # Optional. The name of the agent evaluation dataset. - } - - # response body for status code(s): 200 - response == { - "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @overload - async def create_evaluation_dataset( + @distributed_trace_async + async def list_custom_models( self, - body: Optional[IO[bytes]] = None, *, - content_type: str = "application/json", + page: Optional[int] = None, + per_page: Optional[int] = None, + status: str = "STATUS_UNSPECIFIED", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Dataset. + """List Custom Models. - To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. + To list custom models, send a GET request to ``/v2/gen-ai/custom_models``. - :param body: Default value is None. - :type body: IO[bytes] - :keyword content_type: Body Parameter content-type. Content type parameter for binary body. - Default value is "application/json". - :paramtype content_type: str + :keyword page: Page number for pagination. Default value is None. + :paramtype page: int + :keyword per_page: Number of items per page. Default value is None. + :paramtype per_page: int + :keyword status: Filter by model status. Known values are: "STATUS_UNSPECIFIED", + "STATUS_IMPORTING", "STATUS_READY", "STATUS_FAILED", and "STATUS_DELETED". Default value is + "STATUS_UNSPECIFIED". + :paramtype status: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -229186,57 +238528,118 @@ async def create_evaluation_dataset( # response body for status code(s): 200 response == { - "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @distributed_trace_async - async def create_evaluation_dataset( - self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Create Evaluation Dataset. - - To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. - - :param body: Is either a JSON type or a IO[bytes] type. Default value is None. - :type body: JSON or IO[bytes] - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value - is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: - "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", and - "EVALUATION_DATASET_TYPE_NON_ADK". - "file_upload_dataset": { - "original_file_name": "str", # Optional. The original file name. - "size_in_bytes": "str", # Optional. The size of the file in bytes. - "stored_object_key": "str" # Optional. The object key the file was - stored as. + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } }, - "name": "str" # Optional. The name of the agent evaluation dataset. - } - - # response body for status code(s): 200 - response == { - "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. + "max_threshold": 0, # Optional. Maximum number of custom models allowed for + this team's tier. + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "models": [ + { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 + timestamp indicating when the dedicated inference deployment was + created. + "endpoints": { + "private_endpoint_fqdn": "str", # + Optional. Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # + Optional. Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier + (UUID) of the dedicated inference deployment. + "name": "str", # Optional. Human-readable + name of the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of + the region where the dedicated inference deployment is running + (e.g. "atl1"). + "state": "str", # Optional. Current + lifecycle state of the dedicated inference deployment (e.g. + "ACTIVE", "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 + timestamp indicating when the dedicated inference deployment was + last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from + the model repository. + "context_length": 0, # Optional. Maximum context length + supported by the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly + cost in dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp + when the model was created. + "description": "str", # Optional. Description of the custom + model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., + text, image). + ], + "license": "str", # Optional. License under which the model + is distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., + text, image). + ], + "parameters": "str", # Optional. Number of parameters in the + model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # + Optional. Default value is "ACCESS_TYPE_UNSPECIFIED". Access level + required for the model repository. Known values are: + "ACCESS_TYPE_UNSPECIFIED", "ACCESS_TYPE_PUBLIC", + "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of + the model version. + "hf_token": "str", # Optional. User-provided + HuggingFace token for gated/private models (not persisted in + source_ref). + "prefix": "str", # Optional. Object prefix path in + the bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. + Default value is "SOURCE_TYPE_UNSPECIFIED". Source from which the model + was imported. Known values are: "SOURCE_TYPE_UNSPECIFIED", + "SOURCE_TYPE_HUGGINGFACE", "SOURCE_TYPE_SPACES_BUCKET", + "SOURCE_TYPE_SDK_UPLOAD", and "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. + Known values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", + "STATUS_READY", "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces + bucket where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model + files in bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp + when the model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom + model. + } + ] } # response body for status code(s): 404 response == { @@ -229263,29 +238666,15 @@ async def create_evaluation_dataset( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} - content_type: Optional[str] = kwargs.pop( - "content_type", _headers.pop("Content-Type", None) - ) cls: ClsType[JSON] = kwargs.pop("cls", None) - content_type = content_type or "application/json" - _json = None - _content = None - if isinstance(body, (IOBase, bytes)): - _content = body - else: - if body is not None: - _json = body - else: - _json = None - - _request = build_genai_create_evaluation_dataset_request( - content_type=content_type, - json=_json, - content=_content, + _request = build_genai_list_custom_models_request( + page=page, + per_page=per_page, + status=status, headers=_headers, params=_params, ) @@ -229345,7 +238734,7 @@ async def create_evaluation_dataset( return cast(JSON, deserialized) # type: ignore @overload - async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + async def import_custom_model( self, body: Optional[JSON] = None, *, @@ -229353,10 +238742,9 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Presigned URLs for Evaluation Dataset File Upload. + """Import Custom Model. - To create presigned URLs for evaluation dataset file upload, send a POST request to - ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + To import a custom model, send a POST request to ``/v2/gen-ai/custom_models/import``. :param body: Default value is None. :type body: JSON @@ -229372,29 +238760,151 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab # JSON input template you can fill out and use as your body input. body = { - "files": [ - { - "file_name": "str", # Optional. Local filename. - "file_size": "str" # Optional. The size of the file in - bytes. - } - ] + "accept_terms_and_conditions": bool, # Optional. Whether the caller accepts + the terms and conditions for importing this model. + "description": "str", # Optional. Description of the model. + "name": "str", # Optional. Name for the imported model. + "preferred_gpu_region": "str", # Optional. Preferred GPU region for + deployment. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. Default value + is "ACCESS_TYPE_UNSPECIFIED". Access level required for the model repository. + Known values are: "ACCESS_TYPE_UNSPECIFIED", "ACCESS_TYPE_PUBLIC", + "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace token for + gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value is + "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known values + are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + } } # response body for status code(s): 200 response == { - "request_id": "str", # Optional. The ID generated for the request for - Presigned URLs. - "uploads": [ + "error": "str", # Optional. + "import_job": { + "bytes_done": "str", # Optional. Bytes imported so far. + "bytes_total": "str", # Optional. Total bytes to import. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp when + the import completed. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + job was created. + "error_message": "str", # Optional. Error message if import failed. + "error_step": "str", # Optional. Step at which the error occurred. + "files_done": 0, # Optional. Number of files imported so far. + "files_total": 0, # Optional. Total number of files to import. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + import started. + "status": "str", # Optional. Current status of the import job. + "uuid": "str" # Optional. Unique identifier for the import job. + }, + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + }, + "validation_steps": [ { - "expires_at": "2020-02-20 00:00:00", # Optional. The time - the url expires at. - "object_key": "str", # Optional. The unique object key to - store the file as. - "original_file_name": "str", # Optional. The original file - name. - "presigned_url": "str" # Optional. The actual presigned URL - the client can use to upload the file directly. + "error": "str", # Optional. Error message if validation + failed. + "name": "str", # Optional. Name of the validation step. + "passed": bool # Optional. Whether the validation step + passed. } ] } @@ -229412,7 +238922,7 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab """ @overload - async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + async def import_custom_model( self, body: Optional[IO[bytes]] = None, *, @@ -229420,10 +238930,9 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Presigned URLs for Evaluation Dataset File Upload. + """Import Custom Model. - To create presigned URLs for evaluation dataset file upload, send a POST request to - ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + To import a custom model, send a POST request to ``/v2/gen-ai/custom_models/import``. :param body: Default value is None. :type body: IO[bytes] @@ -229439,18 +238948,117 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab # response body for status code(s): 200 response == { - "request_id": "str", # Optional. The ID generated for the request for - Presigned URLs. - "uploads": [ + "error": "str", # Optional. + "import_job": { + "bytes_done": "str", # Optional. Bytes imported so far. + "bytes_total": "str", # Optional. Total bytes to import. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp when + the import completed. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + job was created. + "error_message": "str", # Optional. Error message if import failed. + "error_step": "str", # Optional. Step at which the error occurred. + "files_done": 0, # Optional. Number of files imported so far. + "files_total": 0, # Optional. Total number of files to import. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + import started. + "status": "str", # Optional. Current status of the import job. + "uuid": "str" # Optional. Unique identifier for the import job. + }, + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + }, + "validation_steps": [ { - "expires_at": "2020-02-20 00:00:00", # Optional. The time - the url expires at. - "object_key": "str", # Optional. The unique object key to - store the file as. - "original_file_name": "str", # Optional. The original file - name. - "presigned_url": "str" # Optional. The actual presigned URL - the client can use to upload the file directly. + "error": "str", # Optional. Error message if validation + failed. + "name": "str", # Optional. Name of the validation step. + "passed": bool # Optional. Whether the validation step + passed. } ] } @@ -229468,14 +239076,13 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab """ @distributed_trace_async - async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + async def import_custom_model( self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Presigned URLs for Evaluation Dataset File Upload. + """Import Custom Model. - To create presigned URLs for evaluation dataset file upload, send a POST request to - ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + To import a custom model, send a POST request to ``/v2/gen-ai/custom_models/import``. :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] @@ -229488,29 +239095,151 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab # JSON input template you can fill out and use as your body input. body = { - "files": [ - { - "file_name": "str", # Optional. Local filename. - "file_size": "str" # Optional. The size of the file in - bytes. - } - ] + "accept_terms_and_conditions": bool, # Optional. Whether the caller accepts + the terms and conditions for importing this model. + "description": "str", # Optional. Description of the model. + "name": "str", # Optional. Name for the imported model. + "preferred_gpu_region": "str", # Optional. Preferred GPU region for + deployment. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. Default value + is "ACCESS_TYPE_UNSPECIFIED". Access level required for the model repository. + Known values are: "ACCESS_TYPE_UNSPECIFIED", "ACCESS_TYPE_PUBLIC", + "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace token for + gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value is + "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known values + are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + } } # response body for status code(s): 200 response == { - "request_id": "str", # Optional. The ID generated for the request for - Presigned URLs. - "uploads": [ + "error": "str", # Optional. + "import_job": { + "bytes_done": "str", # Optional. Bytes imported so far. + "bytes_total": "str", # Optional. Total bytes to import. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp when + the import completed. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + job was created. + "error_message": "str", # Optional. Error message if import failed. + "error_step": "str", # Optional. Step at which the error occurred. + "files_done": 0, # Optional. Number of files imported so far. + "files_total": 0, # Optional. Total number of files to import. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + import started. + "status": "str", # Optional. Current status of the import job. + "uuid": "str" # Optional. Unique identifier for the import job. + }, + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + }, + "validation_steps": [ { - "expires_at": "2020-02-20 00:00:00", # Optional. The time - the url expires at. - "object_key": "str", # Optional. The unique object key to - store the file as. - "original_file_name": "str", # Optional. The original file - name. - "presigned_url": "str" # Optional. The actual presigned URL - the client can use to upload the file directly. + "error": "str", # Optional. Error message if validation + failed. + "name": "str", # Optional. Name of the validation step. + "passed": bool # Optional. Whether the validation step + passed. } ] } @@ -229558,14 +239287,12 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab else: _json = None - _request = ( - build_genai_create_evaluation_dataset_file_upload_presigned_urls_request( - content_type=content_type, - json=_json, - content=_content, - headers=_headers, - params=_params, - ) + _request = build_genai_import_custom_model_request( + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, ) _request.url = self._client.format_url(_request.url) @@ -229623,12 +239350,14 @@ async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disab return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def list_evaluation_metrics(self, **kwargs: Any) -> JSON: + async def delete_custom_model(self, uuid: str, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Evaluation Metrics. + """Delete Custom Model. - To list all evaluation metrics, send a GET request to ``/v2/gen-ai/evaluation_metrics``. + To delete a custom model, send a DELETE request to ``/v2/genai/custom_models/{uuid}``. + :param uuid: UUID of the custom model to delete. Required. + :type uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -229638,34 +239367,11 @@ async def list_evaluation_metrics(self, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "metrics": [ - { - "category": "METRIC_CATEGORY_UNSPECIFIED", # Optional. - Default value is "METRIC_CATEGORY_UNSPECIFIED". Known values are: - "METRIC_CATEGORY_UNSPECIFIED", "METRIC_CATEGORY_CORRECTNESS", - "METRIC_CATEGORY_USER_OUTCOMES", "METRIC_CATEGORY_SAFETY_AND_SECURITY", - "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". - "description": "str", # Optional. - "inverted": bool, # Optional. If true, the metric is - inverted, meaning that a lower value is better. - "is_metric_goal": bool, # Optional. - "metric_name": "str", # Optional. - "metric_rank": 0, # Optional. - "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. - Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". - "metric_uuid": "str", # Optional. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "range_max": 0.0, # Optional. The maximum value for the - metric. - "range_min": 0.0 # Optional. The minimum value for the - metric. - } - ] + "error": "str", # Optional. Error message if deletion failed. + "status": "DELETE_CUSTOM_MODEL_STATUS_UNSPECIFIED" # Optional. Default value + is "DELETE_CUSTOM_MODEL_STATUS_UNSPECIFIED". Status of delete operation. Known + values are: "DELETE_CUSTOM_MODEL_STATUS_UNSPECIFIED", + "DELETE_CUSTOM_MODEL_STATUS_SUCCESS", and "DELETE_CUSTOM_MODEL_STATUS_FAIL". } # response body for status code(s): 404 response == { @@ -229697,7 +239403,8 @@ async def list_evaluation_metrics(self, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_evaluation_metrics_request( + _request = build_genai_delete_custom_model_request( + uuid=uuid, headers=_headers, params=_params, ) @@ -229757,18 +239464,22 @@ async def list_evaluation_metrics(self, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - async def run_evaluation_test_case( + async def update_custom_model_metadata( self, + uuid: str, body: Optional[JSON] = None, *, content_type: str = "application/json", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Run an Evaluation Test Case. + """Update Custom Model Metadata. - To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. + To update custom model metadata, send a PATCH request to + ``/v2/gen-ai/custom_models/{uuid}/metadata``. + :param uuid: UUID of the custom model to update. Required. + :type uuid: str :param body: Default value is None. :type body: JSON :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. @@ -229783,23 +239494,105 @@ async def run_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_deployment_names": [ - "str" # Optional. Agent deployment names to run the test case - against (ADK agent workspaces). - ], - "agent_uuids": [ - "str" # Optional. Agent UUIDs to run the test case against (legacy - agents). - ], - "run_name": "str", # Optional. The name of the run. - "test_case_uuid": "str" # Optional. Test-case UUID to run. + "description": "str", # Optional. + "name": "str", # Optional. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "uuid": "str" # Optional. UUID of the custom model to update. } # response body for status code(s): 200 response == { - "evaluation_run_uuids": [ - "str" # Optional. - ] + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + } } # response body for status code(s): 404 response == { @@ -229815,18 +239608,22 @@ async def run_evaluation_test_case( """ @overload - async def run_evaluation_test_case( + async def update_custom_model_metadata( self, + uuid: str, body: Optional[IO[bytes]] = None, *, content_type: str = "application/json", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Run an Evaluation Test Case. + """Update Custom Model Metadata. - To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. + To update custom model metadata, send a PATCH request to + ``/v2/gen-ai/custom_models/{uuid}/metadata``. + :param uuid: UUID of the custom model to update. Required. + :type uuid: str :param body: Default value is None. :type body: IO[bytes] :keyword content_type: Body Parameter content-type. Content type parameter for binary body. @@ -229841,9 +239638,93 @@ async def run_evaluation_test_case( # response body for status code(s): 200 response == { - "evaluation_run_uuids": [ - "str" # Optional. - ] + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + } } # response body for status code(s): 404 response == { @@ -229859,14 +239740,17 @@ async def run_evaluation_test_case( """ @distributed_trace_async - async def run_evaluation_test_case( - self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + async def update_custom_model_metadata( + self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Run an Evaluation Test Case. + """Update Custom Model Metadata. - To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. + To update custom model metadata, send a PATCH request to + ``/v2/gen-ai/custom_models/{uuid}/metadata``. + :param uuid: UUID of the custom model to update. Required. + :type uuid: str :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] :return: JSON object @@ -229878,23 +239762,105 @@ async def run_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_deployment_names": [ - "str" # Optional. Agent deployment names to run the test case - against (ADK agent workspaces). - ], - "agent_uuids": [ - "str" # Optional. Agent UUIDs to run the test case against (legacy - agents). - ], - "run_name": "str", # Optional. The name of the run. - "test_case_uuid": "str" # Optional. Test-case UUID to run. + "description": "str", # Optional. + "name": "str", # Optional. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "uuid": "str" # Optional. UUID of the custom model to update. } # response body for status code(s): 200 response == { - "evaluation_run_uuids": [ - "str" # Optional. - ] + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + } } # response body for status code(s): 404 response == { @@ -229940,7 +239906,8 @@ async def run_evaluation_test_case( else: _json = None - _request = build_genai_run_evaluation_test_case_request( + _request = build_genai_update_custom_model_metadata_request( + uuid=uuid, content_type=content_type, json=_json, content=_content, @@ -230002,16 +239969,116 @@ async def run_evaluation_test_case( return cast(JSON, deserialized) # type: ignore + @overload + async def create_evaluation_dataset( + self, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create Evaluation Dataset. + + To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. + + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and "EVALUATION_DATASET_TYPE_MODEL". + "file_upload_dataset": { + "original_file_name": "str", # Optional. The original file name. + "size_in_bytes": "str", # Optional. The size of the file in bytes. + "stored_object_key": "str" # Optional. The object key the file was + stored as. + }, + "name": "str" # Optional. The name of the agent evaluation dataset. + } + + # response body for status code(s): 200 + response == { + "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + async def create_evaluation_dataset( + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create Evaluation Dataset. + + To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + @distributed_trace_async - async def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> JSON: + async def create_evaluation_dataset( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: # pylint: disable=line-too-long - """Retrieve Information About an Existing Evaluation Run. + """Create Evaluation Dataset. - To retrive information about an existing evaluation run, send a GET request to - ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}``. + To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. - :param evaluation_run_uuid: Evaluation run UUID. Required. - :type evaluation_run_uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -230019,73 +240086,24 @@ async def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> J Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and "EVALUATION_DATASET_TYPE_MODEL". + "file_upload_dataset": { + "original_file_name": "str", # Optional. The original file name. + "size_in_bytes": "str", # Optional. The size of the file in bytes. + "stored_object_key": "str" # Optional. The object key the file was + stored as. + }, + "name": "str" # Optional. The name of the agent evaluation dataset. + } + # response body for status code(s): 200 response == { - "evaluation_run": { - "agent_deleted": bool, # Optional. Whether agent is deleted. - "agent_deployment_name": "str", # Optional. The agent deployment - name. - "agent_name": "str", # Optional. Agent name. - "agent_uuid": "str", # Optional. Agent UUID. - "agent_version_hash": "str", # Optional. Version hash. - "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. - "created_by_user_email": "str", # Optional. - "created_by_user_id": "str", # Optional. - "error_description": "str", # Optional. The error description. - "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. - "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation - test case workspace uuid. - "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. - "pass_status": bool, # Optional. The pass status of the evaluation - run based on the star metric. - "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. - "run_level_metric_results": [ - { - "error_description": "str", # Optional. Error - description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", - # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known - values are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the - metric as a number. - "reasoning": "str", # Optional. Reasoning of the - metric result. - "string_value": "str" # Optional. The value of the - metric as a string. - } - ], - "run_name": "str", # Optional. Run name. - "star_metric_result": { - "error_description": "str", # Optional. Error description if - the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the metric as - a number. - "reasoning": "str", # Optional. Reasoning of the metric - result. - "string_value": "str" # Optional. The value of the metric as - a string. - }, - "started_at": "2020-02-20 00:00:00", # Optional. Run start time. - "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default - value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known - values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", - "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", - "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", - "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and - "EVALUATION_RUN_FAILED". - "test_case_description": "str", # Optional. Test case description. - "test_case_name": "str", # Optional. Test case name. - "test_case_uuid": "str", # Optional. Test-case UUID. - "test_case_version": 0 # Optional. Test-case-version. - } + "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. } # response body for status code(s): 404 response == { @@ -230112,13 +240130,29 @@ async def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> J } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = kwargs.pop("headers", {}) or {} + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_run_request( - evaluation_run_uuid=evaluation_run_uuid, + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_create_evaluation_dataset_request( + content_type=content_type, + json=_json, + content=_content, headers=_headers, params=_params, ) @@ -230177,27 +240211,25 @@ async def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> J return cast(JSON, deserialized) # type: ignore - @distributed_trace_async - async def get_evaluation_run_results( + @overload + async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long self, - evaluation_run_uuid: str, + body: Optional[JSON] = None, *, - page: Optional[int] = None, - per_page: Optional[int] = None, + content_type: str = "application/json", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Retrieve Results of an Evaluation Run. + """Create Presigned URLs for Evaluation Dataset File Upload. - To retrieve results of an evaluation run, send a GET request to - ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results``. + To create presigned URLs for evaluation dataset file upload, send a POST request to + ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. - :param evaluation_run_uuid: Evaluation run UUID. Required. - :type evaluation_run_uuid: str - :keyword page: Page number. Default value is None. - :paramtype page: int - :keyword per_page: Items per page. Default value is None. - :paramtype per_page: int + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -230205,184 +240237,147 @@ async def get_evaluation_run_results( Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + # response body for status code(s): 200 response == { - "evaluation_run": { - "agent_deleted": bool, # Optional. Whether agent is deleted. - "agent_deployment_name": "str", # Optional. The agent deployment - name. - "agent_name": "str", # Optional. Agent name. - "agent_uuid": "str", # Optional. Agent UUID. - "agent_version_hash": "str", # Optional. Version hash. - "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. - "created_by_user_email": "str", # Optional. - "created_by_user_id": "str", # Optional. - "error_description": "str", # Optional. The error description. - "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. - "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation - test case workspace uuid. - "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. - "pass_status": bool, # Optional. The pass status of the evaluation - run based on the star metric. - "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. - "run_level_metric_results": [ - { - "error_description": "str", # Optional. Error - description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", - # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known - values are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the - metric as a number. - "reasoning": "str", # Optional. Reasoning of the - metric result. - "string_value": "str" # Optional. The value of the - metric as a string. - } - ], - "run_name": "str", # Optional. Run name. - "star_metric_result": { - "error_description": "str", # Optional. Error description if - the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the metric as - a number. - "reasoning": "str", # Optional. Reasoning of the metric - result. - "string_value": "str" # Optional. The value of the metric as - a string. - }, - "started_at": "2020-02-20 00:00:00", # Optional. Run start time. - "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default - value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known - values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", - "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", - "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", - "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and - "EVALUATION_RUN_FAILED". - "test_case_description": "str", # Optional. Test case description. - "test_case_name": "str", # Optional. Test case name. - "test_case_uuid": "str", # Optional. Test-case UUID. - "test_case_version": 0 # Optional. Test-case-version. - }, - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. - "total": 0 # Optional. Total amount of items over all pages. - }, - "prompts": [ + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Evaluation Dataset File Upload. + + To create presigned URLs for evaluation dataset file upload, send a POST request to + ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ { - "evaluation_trace_spans": [ - { - "created_at": "2020-02-20 00:00:00", # - Optional. When the span was created. - "input": {}, # Optional. Input data for the - span (flexible structure - can be messages array, string, etc.). - "name": "str", # Optional. Name/identifier - for the span. - "output": {}, # Optional. Output data from - the span (flexible structure - can be message, string, etc.). - "retriever_chunks": [ - { - "chunk_usage_pct": 0.0, # - Optional. The usage percentage of the chunk. - "chunk_used": bool, # - Optional. Indicates if the chunk was used in the prompt. - "index_uuid": "str", # - Optional. The index uuid (Knowledge Base) of the chunk. - "source_name": "str", # - Optional. The source name for the chunk, e.g., the file - name or document title. - "text": "str" # Optional. - Text content of the chunk. - } - ], - "span_level_metric_results": [ - { - "error_description": "str", - # Optional. Error description if the metric could not be - calculated. - "metric_name": "str", # - Optional. Metric name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default - value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", - and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # - Optional. The value of the metric as a number. - "reasoning": "str", # - Optional. Reasoning of the metric result. - "string_value": "str" # - Optional. The value of the metric as a string. - } - ], - "type": "TRACE_SPAN_TYPE_UNKNOWN" # - Optional. Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of - spans in a trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", - "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", and - "TRACE_SPAN_TYPE_TOOL". - } - ], - "ground_truth": "str", # Optional. The ground truth for the - prompt. - "input": "str", # Optional. The prompt level results. - "input_tokens": "str", # Optional. The number of input - tokens used in the prompt. - "output": "str", # Optional. The prompt level results. - "output_tokens": "str", # Optional. The number of output - tokens used in the prompt. - "prompt_chunks": [ - { - "chunk_usage_pct": 0.0, # Optional. The - usage percentage of the chunk. - "chunk_used": bool, # Optional. Indicates if - the chunk was used in the prompt. - "index_uuid": "str", # Optional. The index - uuid (Knowledge Base) of the chunk. - "source_name": "str", # Optional. The source - name for the chunk, e.g., the file name or document title. - "text": "str" # Optional. Text content of - the chunk. - } - ], - "prompt_id": 0, # Optional. Prompt ID. - "prompt_level_metric_results": [ - { - "error_description": "str", # Optional. - Error description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric - name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is - "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value - of the metric as a number. - "reasoning": "str", # Optional. Reasoning of - the metric result. - "string_value": "str" # Optional. The value - of the metric as a string. - } - ], - "trace_id": "str" # Optional. The trace id for the prompt. + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace_async + async def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Evaluation Dataset File Upload. + + To create presigned URLs for evaluation dataset file upload, send a POST request to + ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. } ] } @@ -230411,17 +240406,33 @@ async def get_evaluation_run_results( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = kwargs.pop("headers", {}) or {} + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_run_results_request( - evaluation_run_uuid=evaluation_run_uuid, - page=page, - per_page=per_page, - headers=_headers, - params=_params, + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = ( + build_genai_create_evaluation_dataset_file_upload_presigned_urls_request( + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) ) _request.url = self._client.format_url(_request.url) @@ -230479,19 +240490,17 @@ async def get_evaluation_run_results( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def get_evaluation_run_prompt_results( - self, evaluation_run_uuid: str, prompt_id: int, **kwargs: Any + async def get_evaluation_dataset_download_url( + self, dataset_uuid: str, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Retrieve Results of an Evaluation Run Prompt. + """Get Download URL for Evaluation Dataset. - To retrieve results of an evaluation run, send a GET request to - ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results/{prompt_id}``. + To get a presigned download URL for an evaluation dataset, send a GET request to + ``/v2/genai/evaluation_datasets/{dataset_uuid}/download_url``. - :param evaluation_run_uuid: Evaluation run UUID. Required. - :type evaluation_run_uuid: str - :param prompt_id: Prompt ID to get results for. Required. - :type prompt_id: int + :param dataset_uuid: UUID of the evaluation dataset. Required. + :type dataset_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -230501,101 +240510,9 @@ async def get_evaluation_run_prompt_results( # response body for status code(s): 200 response == { - "prompt": { - "evaluation_trace_spans": [ - { - "created_at": "2020-02-20 00:00:00", # Optional. - When the span was created. - "input": {}, # Optional. Input data for the span - (flexible structure - can be messages array, string, etc.). - "name": "str", # Optional. Name/identifier for the - span. - "output": {}, # Optional. Output data from the span - (flexible structure - can be message, string, etc.). - "retriever_chunks": [ - { - "chunk_usage_pct": 0.0, # Optional. - The usage percentage of the chunk. - "chunk_used": bool, # Optional. - Indicates if the chunk was used in the prompt. - "index_uuid": "str", # Optional. The - index uuid (Knowledge Base) of the chunk. - "source_name": "str", # Optional. - The source name for the chunk, e.g., the file name or - document title. - "text": "str" # Optional. Text - content of the chunk. - } - ], - "span_level_metric_results": [ - { - "error_description": "str", # - Optional. Error description if the metric could not be - calculated. - "metric_name": "str", # Optional. - Metric name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value - is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The - value of the metric as a number. - "reasoning": "str", # Optional. - Reasoning of the metric result. - "string_value": "str" # Optional. - The value of the metric as a string. - } - ], - "type": "TRACE_SPAN_TYPE_UNKNOWN" # Optional. - Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of spans in a - trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", - "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", and - "TRACE_SPAN_TYPE_TOOL". - } - ], - "ground_truth": "str", # Optional. The ground truth for the prompt. - "input": "str", # Optional. - "input_tokens": "str", # Optional. The number of input tokens used - in the prompt. - "output": "str", # Optional. - "output_tokens": "str", # Optional. The number of output tokens used - in the prompt. - "prompt_chunks": [ - { - "chunk_usage_pct": 0.0, # Optional. The usage - percentage of the chunk. - "chunk_used": bool, # Optional. Indicates if the - chunk was used in the prompt. - "index_uuid": "str", # Optional. The index uuid - (Knowledge Base) of the chunk. - "source_name": "str", # Optional. The source name - for the chunk, e.g., the file name or document title. - "text": "str" # Optional. Text content of the chunk. - } - ], - "prompt_id": 0, # Optional. Prompt ID. - "prompt_level_metric_results": [ - { - "error_description": "str", # Optional. Error - description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", - # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known - values are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the - metric as a number. - "reasoning": "str", # Optional. Reasoning of the - metric result. - "string_value": "str" # Optional. The value of the - metric as a string. - } - ], - "trace_id": "str" # Optional. The trace id for the prompt. - } + "download_url": "str", # Optional. The presigned URL to download the dataset + file. + "expires_at": "2020-02-20 00:00:00" # Optional. The time the URL expires at. } # response body for status code(s): 404 response == { @@ -230627,9 +240544,8 @@ async def get_evaluation_run_prompt_results( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_run_prompt_results_request( - evaluation_run_uuid=evaluation_run_uuid, - prompt_id=prompt_id, + _request = build_genai_get_evaluation_dataset_download_url_request( + dataset_uuid=dataset_uuid, headers=_headers, params=_params, ) @@ -230689,11 +240605,11 @@ async def get_evaluation_run_prompt_results( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: + async def list_evaluation_metrics(self, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Evaluation Test Cases. + """List Evaluation Metrics. - To list all evaluation test cases, send a GET request to ``/v2/gen-ai/evaluation_test_cases``. + To list all evaluation metrics, send a GET request to ``/v2/gen-ai/evaluation_metrics``. :return: JSON object :rtype: JSON @@ -230704,120 +240620,40 @@ async def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "evaluation_test_cases": [ + "metrics": [ { - "archived_at": "2020-02-20 00:00:00", # Optional. - Alternative way of authentication for internal usage only - should not be - exposed to public api. - "created_at": "2020-02-20 00:00:00", # Optional. Alternative - way of authentication for internal usage only - should not be exposed to - public api. - "created_by_user_email": "str", # Optional. Alternative way - of authentication for internal usage only - should not be exposed to - public api. - "created_by_user_id": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "dataset": { - "created_at": "2020-02-20 00:00:00", # Optional. - Time created at. - "dataset_name": "str", # Optional. Name of the - dataset. - "dataset_uuid": "str", # Optional. UUID of the - dataset. - "file_size": "str", # Optional. The size of the - dataset uploaded file in bytes. - "has_ground_truth": bool, # Optional. Does the - dataset have a ground truth column?. - "row_count": 0 # Optional. Number of rows in the - dataset. - }, - "dataset_name": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "dataset_uuid": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "description": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "latest_version_number_of_runs": 0, # Optional. Alternative - way of authentication for internal usage only - should not be exposed to - public api. - "metrics": [ - { - "category": "METRIC_CATEGORY_UNSPECIFIED", # - Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known - values are: "METRIC_CATEGORY_UNSPECIFIED", - "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", - "METRIC_CATEGORY_SAFETY_AND_SECURITY", - "METRIC_CATEGORY_CONTEXT_QUALITY", and - "METRIC_CATEGORY_MODEL_FIT". - "description": "str", # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "inverted": bool, # Optional. If true, the - metric is inverted, meaning that a lower value is better. - "is_metric_goal": bool, # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "metric_name": "str", # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "metric_rank": 0, # Optional. Alternative - way of authentication for internal usage only - should not be - exposed to public api. - "metric_type": "METRIC_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known - values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and "METRIC_TYPE_RAG_AND_TOOL". - "metric_uuid": "str", # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is - "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "range_max": 0.0, # Optional. The maximum - value for the metric. - "range_min": 0.0 # Optional. The minimum - value for the metric. - } - ], - "name": "str", # Optional. Alternative way of authentication - for internal usage only - should not be exposed to public api. - "star_metric": { - "metric_uuid": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to - public api. - "name": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to - public api. - "success_threshold": 0.0, # Optional. The success - threshold for the star metric. This is a value that the metric must - reach to be considered successful. - "success_threshold_pct": 0 # Optional. The success - threshold for the star metric. This is a percentage value between 0 - and 100. - }, - "test_case_uuid": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "total_runs": 0, # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "updated_at": "2020-02-20 00:00:00", # Optional. Alternative - way of authentication for internal usage only - should not be exposed to - public api. - "updated_by_user_email": "str", # Optional. Alternative way - of authentication for internal usage only - should not be exposed to - public api. - "updated_by_user_id": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "version": 0 # Optional. Alternative way of authentication - for internal usage only - should not be exposed to public api. + "category": "METRIC_CATEGORY_UNSPECIFIED", # Optional. + Default value is "METRIC_CATEGORY_UNSPECIFIED". Known values are: + "METRIC_CATEGORY_UNSPECIFIED", "METRIC_CATEGORY_CORRECTNESS", + "METRIC_CATEGORY_USER_OUTCOMES", "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", # + Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. + "metric_name": "str", # Optional. + "metric_rank": 0, # Optional. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. + Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for the + metric. + "range_min": 0.0 # Optional. The minimum value for the + metric. } ] } @@ -230851,7 +240687,7 @@ async def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_evaluation_test_cases_request( + _request = build_genai_list_evaluation_metrics_request( headers=_headers, params=_params, ) @@ -230911,7 +240747,7 @@ async def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - async def create_evaluation_test_case( + async def run_evaluation_test_case( self, body: Optional[JSON] = None, *, @@ -230919,9 +240755,9 @@ async def create_evaluation_test_case( **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Test Case. + """Run an Evaluation Test Case. - To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. + To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. :param body: Default value is None. :type body: JSON @@ -230937,29 +240773,23 @@ async def create_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_workspace_name": "str", # Optional. - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": [ - "str" # Optional. Full metric list to use for evaluation test case. + "agent_deployment_names": [ + "str" # Optional. Agent deployment names to run the test case + against (ADK agent workspaces). ], - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "workspace_uuid": "str" # Optional. The workspace uuid. + "agent_uuids": [ + "str" # Optional. Agent UUIDs to run the test case against (legacy + agents). + ], + "run_name": "str", # Optional. The name of the run. + "test_case_uuid": "str" # Optional. Test-case UUID to run. } # response body for status code(s): 200 response == { - "test_case_uuid": "str" # Optional. Test"u2011case UUID. + "evaluation_run_uuids": [ + "str" # Optional. + ] } # response body for status code(s): 404 response == { @@ -230975,7 +240805,7 @@ async def create_evaluation_test_case( """ @overload - async def create_evaluation_test_case( + async def run_evaluation_test_case( self, body: Optional[IO[bytes]] = None, *, @@ -230983,9 +240813,9 @@ async def create_evaluation_test_case( **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Test Case. + """Run an Evaluation Test Case. - To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. + To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. :param body: Default value is None. :type body: IO[bytes] @@ -231001,7 +240831,9 @@ async def create_evaluation_test_case( # response body for status code(s): 200 response == { - "test_case_uuid": "str" # Optional. Test"u2011case UUID. + "evaluation_run_uuids": [ + "str" # Optional. + ] } # response body for status code(s): 404 response == { @@ -231017,13 +240849,13 @@ async def create_evaluation_test_case( """ @distributed_trace_async - async def create_evaluation_test_case( + async def run_evaluation_test_case( self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Test Case. + """Run an Evaluation Test Case. - To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. + To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] @@ -231036,29 +240868,23 @@ async def create_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_workspace_name": "str", # Optional. - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": [ - "str" # Optional. Full metric list to use for evaluation test case. + "agent_deployment_names": [ + "str" # Optional. Agent deployment names to run the test case + against (ADK agent workspaces). ], - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "workspace_uuid": "str" # Optional. The workspace uuid. + "agent_uuids": [ + "str" # Optional. Agent UUIDs to run the test case against (legacy + agents). + ], + "run_name": "str", # Optional. The name of the run. + "test_case_uuid": "str" # Optional. Test-case UUID to run. } # response body for status code(s): 200 response == { - "test_case_uuid": "str" # Optional. Test"u2011case UUID. + "evaluation_run_uuids": [ + "str" # Optional. + ] } # response body for status code(s): 404 response == { @@ -231104,7 +240930,7 @@ async def create_evaluation_test_case( else: _json = None - _request = build_genai_create_evaluation_test_case_request( + _request = build_genai_run_evaluation_test_case_request( content_type=content_type, json=_json, content=_content, @@ -231167,23 +240993,15 @@ async def create_evaluation_test_case( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def list_evaluation_runs_by_test_case( - self, - evaluation_test_case_uuid: str, - *, - evaluation_test_case_version: Optional[int] = None, - **kwargs: Any - ) -> JSON: + async def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Evaluation Runs by Test Case. + """Retrieve Information About an Existing Evaluation Run. - To list all evaluation runs by test case, send a GET request to - ``/v2/gen-ai/evaluation_test_cases/{evaluation_test_case_uuid}/evaluation_runs``. + To retrive information about an existing evaluation run, send a GET request to + ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}``. - :param evaluation_test_case_uuid: Evaluation run UUID. Required. - :type evaluation_test_case_uuid: str - :keyword evaluation_test_case_version: Version of the test case. Default value is None. - :paramtype evaluation_test_case_version: int + :param evaluation_run_uuid: Evaluation run UUID. Required. + :type evaluation_run_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -231193,53 +241011,26 @@ async def list_evaluation_runs_by_test_case( # response body for status code(s): 200 response == { - "evaluation_runs": [ - { - "agent_deleted": bool, # Optional. Whether agent is deleted. - "agent_deployment_name": "str", # Optional. The agent - deployment name. - "agent_name": "str", # Optional. Agent name. - "agent_uuid": "str", # Optional. Agent UUID. - "agent_version_hash": "str", # Optional. Version hash. - "agent_workspace_uuid": "str", # Optional. Agent workspace - uuid. - "created_by_user_email": "str", # Optional. List of - evaluation runs. - "created_by_user_id": "str", # Optional. List of evaluation - runs. - "error_description": "str", # Optional. The error - description. - "evaluation_run_uuid": "str", # Optional. Evaluation run - UUID. - "evaluation_test_case_workspace_uuid": "str", # Optional. - Evaluation test case workspace uuid. - "finished_at": "2020-02-20 00:00:00", # Optional. Run end - time. - "pass_status": bool, # Optional. The pass status of the - evaluation run based on the star metric. - "queued_at": "2020-02-20 00:00:00", # Optional. Run queued - time. - "run_level_metric_results": [ - { - "error_description": "str", # Optional. - Error description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric - name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is - "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value - of the metric as a number. - "reasoning": "str", # Optional. Reasoning of - the metric result. - "string_value": "str" # Optional. The value - of the metric as a string. - } - ], - "run_name": "str", # Optional. Run name. - "star_metric_result": { + "evaluation_run": { + "agent_deleted": bool, # Optional. Whether agent is deleted. + "agent_deployment_name": "str", # Optional. The agent deployment + name. + "agent_name": "str", # Optional. Agent name. + "agent_uuid": "str", # Optional. Agent UUID. + "agent_version_hash": "str", # Optional. Version hash. + "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. + "created_by_user_email": "str", # Optional. + "created_by_user_id": "str", # Optional. + "error_description": "str", # Optional. The error description. + "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. + "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation + test case workspace uuid. + "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. + "pass_status": bool, # Optional. The pass status of the evaluation + run based on the star metric. + "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. + "run_level_metric_results": [ + { "error_description": "str", # Optional. Error description if the metric could not be calculated. "metric_name": "str", # Optional. Metric name. @@ -231254,33 +241045,47 @@ async def list_evaluation_runs_by_test_case( metric result. "string_value": "str" # Optional. The value of the metric as a string. - }, - "started_at": "2020-02-20 00:00:00", # Optional. Run start - time. - "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. - Default value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run - Statuses. Known values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", - "EVALUATION_RUN_QUEUED", "EVALUATION_RUN_RUNNING_DATASET", - "EVALUATION_RUN_EVALUATING_RESULTS", "EVALUATION_RUN_CANCELLING", - "EVALUATION_RUN_CANCELLED", "EVALUATION_RUN_SUCCESSFUL", - "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and "EVALUATION_RUN_FAILED". - "test_case_description": "str", # Optional. Test case - description. - "test_case_name": "str", # Optional. Test case name. - "test_case_uuid": "str", # Optional. Test-case UUID. - "test_case_version": 0 # Optional. Test-case-version. - } - ] - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support + } + ], + "run_name": "str", # Optional. Run name. + "star_metric_result": { + "error_description": "str", # Optional. Error description if + the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the metric as + a number. + "reasoning": "str", # Optional. Reasoning of the metric + result. + "string_value": "str" # Optional. The value of the metric as + a string. + }, + "started_at": "2020-02-20 00:00:00", # Optional. Run start time. + "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default + value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known + values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", + "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", + "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", + "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "EVALUATION_RUN_FAILED". + "test_case_description": "str", # Optional. Test case description. + "test_case_name": "str", # Optional. Test case name. + "test_case_uuid": "str", # Optional. Test-case UUID. + "test_case_version": 0 # Optional. Test-case-version. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support tickets to help identify the issue. } """ @@ -231302,9 +241107,8 @@ async def list_evaluation_runs_by_test_case( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_evaluation_runs_by_test_case_request( - evaluation_test_case_uuid=evaluation_test_case_uuid, - evaluation_test_case_version=evaluation_test_case_version, + _request = build_genai_get_evaluation_run_request( + evaluation_run_uuid=evaluation_run_uuid, headers=_headers, params=_params, ) @@ -231364,23 +241168,26 @@ async def list_evaluation_runs_by_test_case( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def get_evaluation_test_case( + async def get_evaluation_run_results( self, - test_case_uuid: str, + evaluation_run_uuid: str, *, - evaluation_test_case_version: Optional[int] = None, + page: Optional[int] = None, + per_page: Optional[int] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Retrieve Information About an Existing Evaluation Test Case. + """Retrieve Results of an Evaluation Run. - To retrive information about an existing evaluation test case, send a GET request to - ``/v2/gen-ai/evaluation_test_case/{test_case_uuid}``. + To retrieve results of an evaluation run, send a GET request to + ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results``. - :param test_case_uuid: The test case uuid to retrieve. Required. - :type test_case_uuid: str - :keyword evaluation_test_case_version: Version of the test case. Default value is None. - :paramtype evaluation_test_case_version: int + :param evaluation_run_uuid: Evaluation run UUID. Required. + :type evaluation_run_uuid: str + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -231390,73 +241197,346 @@ async def get_evaluation_test_case( # response body for status code(s): 200 response == { - "evaluation_test_case": { - "archived_at": "2020-02-20 00:00:00", # Optional. - "created_at": "2020-02-20 00:00:00", # Optional. + "evaluation_run": { + "agent_deleted": bool, # Optional. Whether agent is deleted. + "agent_deployment_name": "str", # Optional. The agent deployment + name. + "agent_name": "str", # Optional. Agent name. + "agent_uuid": "str", # Optional. Agent UUID. + "agent_version_hash": "str", # Optional. Version hash. + "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. "created_by_user_email": "str", # Optional. "created_by_user_id": "str", # Optional. - "dataset": { - "created_at": "2020-02-20 00:00:00", # Optional. Time - created at. - "dataset_name": "str", # Optional. Name of the dataset. - "dataset_uuid": "str", # Optional. UUID of the dataset. - "file_size": "str", # Optional. The size of the dataset - uploaded file in bytes. - "has_ground_truth": bool, # Optional. Does the dataset have - a ground truth column?. - "row_count": 0 # Optional. Number of rows in the dataset. - }, - "dataset_name": "str", # Optional. - "dataset_uuid": "str", # Optional. - "description": "str", # Optional. - "latest_version_number_of_runs": 0, # Optional. - "metrics": [ + "error_description": "str", # Optional. The error description. + "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. + "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation + test case workspace uuid. + "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. + "pass_status": bool, # Optional. The pass status of the evaluation + run based on the star metric. + "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. + "run_level_metric_results": [ { - "category": "METRIC_CATEGORY_UNSPECIFIED", # - Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known - values are: "METRIC_CATEGORY_UNSPECIFIED", - "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", - "METRIC_CATEGORY_SAFETY_AND_SECURITY", - "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". - "description": "str", # Optional. - "inverted": bool, # Optional. If true, the metric is - inverted, meaning that a lower value is better. - "is_metric_goal": bool, # Optional. - "metric_name": "str", # Optional. - "metric_rank": 0, # Optional. - "metric_type": "METRIC_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values - are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". - "metric_uuid": "str", # Optional. + "error_description": "str", # Optional. Error + description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "range_max": 0.0, # Optional. The maximum value for - the metric. - "range_min": 0.0 # Optional. The minimum value for - the metric. + "number_value": 0.0, # Optional. The value of the + metric as a number. + "reasoning": "str", # Optional. Reasoning of the + metric result. + "string_value": "str" # Optional. The value of the + metric as a string. } ], - "name": "str", # Optional. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold - for the star metric. This is a value that the metric must reach to be - considered successful. - "success_threshold_pct": 0 # Optional. The success threshold - for the star metric. This is a percentage value between 0 and 100. + "run_name": "str", # Optional. Run name. + "star_metric_result": { + "error_description": "str", # Optional. Error description if + the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the metric as + a number. + "reasoning": "str", # Optional. Reasoning of the metric + result. + "string_value": "str" # Optional. The value of the metric as + a string. }, - "test_case_uuid": "str", # Optional. - "total_runs": 0, # Optional. - "updated_at": "2020-02-20 00:00:00", # Optional. - "updated_by_user_email": "str", # Optional. - "updated_by_user_id": "str", # Optional. - "version": 0 # Optional. - } + "started_at": "2020-02-20 00:00:00", # Optional. Run start time. + "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default + value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known + values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", + "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", + "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", + "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "EVALUATION_RUN_FAILED". + "test_case_description": "str", # Optional. Test case description. + "test_case_name": "str", # Optional. Test case name. + "test_case_uuid": "str", # Optional. Test-case UUID. + "test_case_version": 0 # Optional. Test-case-version. + }, + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "prompts": [ + { + "evaluation_trace_spans": [ + { + "created_at": "2020-02-20 00:00:00", # + Optional. When the span was created. + "input": {}, # Optional. Input data for the + span (flexible structure - can be messages array, string, etc.). + "name": "str", # Optional. Name/identifier + for the span. + "output": {}, # Optional. Output data from + the span (flexible structure - can be message, string, etc.). + "retriever_chunks": [ + { + "chunk_usage_pct": 0.0, # + Optional. The usage percentage of the chunk. + "chunk_used": bool, # + Optional. Indicates if the chunk was used in the prompt. + "index_uuid": "str", # + Optional. The index uuid (Knowledge Base) of the chunk. + "source_name": "str", # + Optional. The source name for the chunk, e.g., the file + name or document title. + "text": "str" # Optional. + Text content of the chunk. + } + ], + "span_level_metric_results": [ + { + "error_description": "str", + # Optional. Error description if the metric could not be + calculated. + "metric_name": "str", # + Optional. Metric name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default + value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", + and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # + Optional. The value of the metric as a number. + "reasoning": "str", # + Optional. Reasoning of the metric result. + "string_value": "str" # + Optional. The value of the metric as a string. + } + ], + "spans": [ + { + "agent": { + "agent_type": + "AGENT_TYPE_UNSPECIFIED", # Optional. Default value + is "AGENT_TYPE_UNSPECIFIED". Agent span. Known values + are: "AGENT_TYPE_UNSPECIFIED", "AGENT_TYPE_DEFAULT", + "AGENT_TYPE_PLANNER", "AGENT_TYPE_REACT", + "AGENT_TYPE_REFLECTION", "AGENT_TYPE_ROUTER", + "AGENT_TYPE_CLASSIFIER", "AGENT_TYPE_SUPERVISOR", and + "AGENT_TYPE_JUDGE". + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "redacted_input": + "str", # Optional. Child spans - must contain + between 1 and 999 spans Allowed types: agent, llm, + tool, retriever (not workflow). + "redacted_output": + "str", # Optional. Child spans - must contain + between 1 and 999 spans Allowed types: agent, llm, + tool, retriever (not workflow). + "spans": [ + ... + ] + }, + "created_at": "2020-02-20 + 00:00:00", # Optional. When the span was created. + "input": {}, # Optional. + Input data for the span (flexible structure - can be + messages array, string, etc.). + "llm": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "model": "str", # + Optional. LLM span. + "num_input_tokens": + 0, # Optional. LLM span. + "num_output_tokens": + 0, # Optional. LLM span. + "temperature": 0.0, + # Optional. LLM span. + "time_to_first_token_ns": "str", # Optional. LLM + span. + "tools": [ + {} # + Optional. Tool definitions passed to the model. + ], + "total_tokens": 0 # + Optional. LLM span. + }, + "name": "str", # Optional. + Name/identifier for the span. + "output": {}, # Optional. + Output data from the span (flexible structure - can be + message, string, etc.). + "retriever": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + } + }, + "tool": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "tool_call_id": "str" + # Optional. Tool span. + }, + "type": + "TRACE_SPAN_TYPE_UNKNOWN", # Optional. Default value is + "TRACE_SPAN_TYPE_UNKNOWN". Types of spans in a trace. + Known values are: "TRACE_SPAN_TYPE_UNKNOWN", + "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", + "TRACE_SPAN_TYPE_TOOL", "TRACE_SPAN_TYPE_AGENT", and + "TRACE_SPAN_TYPE_WORKFLOW". + "workflow": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "spans": [ + ... + ] + } + } + ], + "type": "TRACE_SPAN_TYPE_UNKNOWN" # + Optional. Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of + spans in a trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", + "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", + "TRACE_SPAN_TYPE_TOOL", "TRACE_SPAN_TYPE_AGENT", and + "TRACE_SPAN_TYPE_WORKFLOW". + } + ], + "ground_truth": "str", # Optional. The ground truth for the + prompt. + "input": "str", # Optional. The prompt level results. + "input_tokens": "str", # Optional. The number of input + tokens used in the prompt. + "output": "str", # Optional. The prompt level results. + "output_tokens": "str", # Optional. The number of output + tokens used in the prompt. + "prompt_chunks": [ + { + "chunk_usage_pct": 0.0, # Optional. The + usage percentage of the chunk. + "chunk_used": bool, # Optional. Indicates if + the chunk was used in the prompt. + "index_uuid": "str", # Optional. The index + uuid (Knowledge Base) of the chunk. + "source_name": "str", # Optional. The source + name for the chunk, e.g., the file name or document title. + "text": "str" # Optional. Text content of + the chunk. + } + ], + "prompt_id": 0, # Optional. Prompt ID. + "prompt_level_metric_results": [ + { + "error_description": "str", # Optional. + Error description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric + name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value + of the metric as a number. + "reasoning": "str", # Optional. Reasoning of + the metric result. + "string_value": "str" # Optional. The value + of the metric as a string. + } + ], + "trace_id": "str" # Optional. The trace id for the prompt. + } + ] } # response body for status code(s): 404 response == { @@ -231488,9 +241568,10 @@ async def get_evaluation_test_case( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_test_case_request( - test_case_uuid=test_case_uuid, - evaluation_test_case_version=evaluation_test_case_version, + _request = build_genai_get_evaluation_run_results_request( + evaluation_run_uuid=evaluation_run_uuid, + page=page, + per_page=per_page, headers=_headers, params=_params, ) @@ -231549,140 +241630,20 @@ async def get_evaluation_test_case( return cast(JSON, deserialized) # type: ignore - @overload - async def update_evaluation_test_case( - self, - test_case_uuid: str, - body: Optional[JSON] = None, - *, - content_type: str = "application/json", - **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Update an Evaluation Test Case. - - To update an evaluation test-case send a PUT request to - ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. - - :param test_case_uuid: Test-case UUID to update. Required. - :type test_case_uuid: str - :param body: Default value is None. - :type body: JSON - :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": { - "metric_uuids": [ - "str" # Optional. - ] - }, - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "test_case_uuid": "str" # Optional. Test-case UUID to update. - } - - # response body for status code(s): 200 - response == { - "test_case_uuid": "str", # Optional. - "version": 0 # Optional. The new verson of the test case. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @overload - async def update_evaluation_test_case( - self, - test_case_uuid: str, - body: Optional[IO[bytes]] = None, - *, - content_type: str = "application/json", - **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Update an Evaluation Test Case. - - To update an evaluation test-case send a PUT request to - ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. - - :param test_case_uuid: Test-case UUID to update. Required. - :type test_case_uuid: str - :param body: Default value is None. - :type body: IO[bytes] - :keyword content_type: Body Parameter content-type. Content type parameter for binary body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "test_case_uuid": "str", # Optional. - "version": 0 # Optional. The new verson of the test case. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - @distributed_trace_async - async def update_evaluation_test_case( - self, - test_case_uuid: str, - body: Optional[Union[JSON, IO[bytes]]] = None, - **kwargs: Any + async def get_evaluation_run_prompt_results( + self, evaluation_run_uuid: str, prompt_id: int, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Update an Evaluation Test Case. + """Retrieve Results of an Evaluation Run Prompt. - To update an evaluation test-case send a PUT request to - ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. + To retrieve results of an evaluation run, send a GET request to + ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results/{prompt_id}``. - :param test_case_uuid: Test-case UUID to update. Required. - :type test_case_uuid: str - :param body: Is either a JSON type or a IO[bytes] type. Default value is None. - :type body: JSON or IO[bytes] + :param evaluation_run_uuid: Evaluation run UUID. Required. + :type evaluation_run_uuid: str + :param prompt_id: Prompt ID to get results for. Required. + :type prompt_id: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -231690,33 +241651,269 @@ async def update_evaluation_test_case( Example: .. code-block:: python - # JSON input template you can fill out and use as your body input. - body = { - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": { - "metric_uuids": [ - "str" # Optional. - ] - }, - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "test_case_uuid": "str" # Optional. Test-case UUID to update. - } - # response body for status code(s): 200 response == { - "test_case_uuid": "str", # Optional. - "version": 0 # Optional. The new verson of the test case. + "prompt": { + "evaluation_trace_spans": [ + { + "created_at": "2020-02-20 00:00:00", # Optional. + When the span was created. + "input": {}, # Optional. Input data for the span + (flexible structure - can be messages array, string, etc.). + "name": "str", # Optional. Name/identifier for the + span. + "output": {}, # Optional. Output data from the span + (flexible structure - can be message, string, etc.). + "retriever_chunks": [ + { + "chunk_usage_pct": 0.0, # Optional. + The usage percentage of the chunk. + "chunk_used": bool, # Optional. + Indicates if the chunk was used in the prompt. + "index_uuid": "str", # Optional. The + index uuid (Knowledge Base) of the chunk. + "source_name": "str", # Optional. + The source name for the chunk, e.g., the file name or + document title. + "text": "str" # Optional. Text + content of the chunk. + } + ], + "span_level_metric_results": [ + { + "error_description": "str", # + Optional. Error description if the metric could not be + calculated. + "metric_name": "str", # Optional. + Metric name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value + is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The + value of the metric as a number. + "reasoning": "str", # Optional. + Reasoning of the metric result. + "string_value": "str" # Optional. + The value of the metric as a string. + } + ], + "spans": [ + { + "agent": { + "agent_type": + "AGENT_TYPE_UNSPECIFIED", # Optional. Default value is + "AGENT_TYPE_UNSPECIFIED". Agent span. Known values are: + "AGENT_TYPE_UNSPECIFIED", "AGENT_TYPE_DEFAULT", + "AGENT_TYPE_PLANNER", "AGENT_TYPE_REACT", + "AGENT_TYPE_REFLECTION", "AGENT_TYPE_ROUTER", + "AGENT_TYPE_CLASSIFIER", "AGENT_TYPE_SUPERVISOR", and + "AGENT_TYPE_JUDGE". + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "redacted_input": "str", # + Optional. Child spans - must contain between 1 and 999 + spans Allowed types: agent, llm, tool, retriever (not + workflow). + "redacted_output": "str", # + Optional. Child spans - must contain between 1 and 999 + spans Allowed types: agent, llm, tool, retriever (not + workflow). + "spans": [ + ... + ] + }, + "created_at": "2020-02-20 00:00:00", + # Optional. When the span was created. + "input": {}, # Optional. Input data + for the span (flexible structure - can be messages array, + string, etc.). + "llm": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "model": "str", # Optional. + LLM span. + "num_input_tokens": 0, # + Optional. LLM span. + "num_output_tokens": 0, # + Optional. LLM span. + "temperature": 0.0, # + Optional. LLM span. + "time_to_first_token_ns": + "str", # Optional. LLM span. + "tools": [ + {} # Optional. Tool + definitions passed to the model. + ], + "total_tokens": 0 # + Optional. LLM span. + }, + "name": "str", # Optional. + Name/identifier for the span. + "output": {}, # Optional. Output + data from the span (flexible structure - can be message, + string, etc.). + "retriever": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + } + }, + "tool": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "tool_call_id": "str" # + Optional. Tool span. + }, + "type": "TRACE_SPAN_TYPE_UNKNOWN", # + Optional. Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types + of spans in a trace. Known values are: + "TRACE_SPAN_TYPE_UNKNOWN", "TRACE_SPAN_TYPE_LLM", + "TRACE_SPAN_TYPE_RETRIEVER", "TRACE_SPAN_TYPE_TOOL", + "TRACE_SPAN_TYPE_AGENT", and "TRACE_SPAN_TYPE_WORKFLOW". + "workflow": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "spans": [ + ... + ] + } + } + ], + "type": "TRACE_SPAN_TYPE_UNKNOWN" # Optional. + Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of spans in a + trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", + "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", + "TRACE_SPAN_TYPE_TOOL", "TRACE_SPAN_TYPE_AGENT", and + "TRACE_SPAN_TYPE_WORKFLOW". + } + ], + "ground_truth": "str", # Optional. The ground truth for the prompt. + "input": "str", # Optional. + "input_tokens": "str", # Optional. The number of input tokens used + in the prompt. + "output": "str", # Optional. + "output_tokens": "str", # Optional. The number of output tokens used + in the prompt. + "prompt_chunks": [ + { + "chunk_usage_pct": 0.0, # Optional. The usage + percentage of the chunk. + "chunk_used": bool, # Optional. Indicates if the + chunk was used in the prompt. + "index_uuid": "str", # Optional. The index uuid + (Knowledge Base) of the chunk. + "source_name": "str", # Optional. The source name + for the chunk, e.g., the file name or document title. + "text": "str" # Optional. Text content of the chunk. + } + ], + "prompt_id": 0, # Optional. Prompt ID. + "prompt_level_metric_results": [ + { + "error_description": "str", # Optional. Error + description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the + metric as a number. + "reasoning": "str", # Optional. Reasoning of the + metric result. + "string_value": "str" # Optional. The value of the + metric as a string. + } + ], + "trace_id": "str" # Optional. The trace id for the prompt. + } } # response body for status code(s): 404 response == { @@ -231743,30 +241940,14 @@ async def update_evaluation_test_case( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} - content_type: Optional[str] = kwargs.pop( - "content_type", _headers.pop("Content-Type", None) - ) cls: ClsType[JSON] = kwargs.pop("cls", None) - content_type = content_type or "application/json" - _json = None - _content = None - if isinstance(body, (IOBase, bytes)): - _content = body - else: - if body is not None: - _json = body - else: - _json = None - - _request = build_genai_update_evaluation_test_case_request( - test_case_uuid=test_case_uuid, - content_type=content_type, - json=_json, - content=_content, + _request = build_genai_get_evaluation_run_prompt_results_request( + evaluation_run_uuid=evaluation_run_uuid, + prompt_id=prompt_id, headers=_headers, params=_params, ) @@ -231826,23 +242007,12 @@ async def update_evaluation_test_case( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def list_indexing_jobs( - self, - *, - page: Optional[int] = None, - per_page: Optional[int] = None, - **kwargs: Any - ) -> JSON: + async def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Indexing Jobs for a Knowledge Base. + """List Evaluation Test Cases. - To list all indexing jobs for a knowledge base, send a GET request to - ``/v2/gen-ai/indexing_jobs``. + To list all evaluation test cases, send a GET request to ``/v2/gen-ai/evaluation_test_cases``. - :keyword page: Page number. Default value is None. - :paramtype page: int - :keyword per_page: Items per page. Default value is None. - :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -231852,97 +242022,136 @@ async def list_indexing_jobs( # response body for status code(s): 200 response == { - "jobs": [ + "evaluation_test_cases": [ { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ + "archived_at": "2020-02-20 00:00:00", # Optional. + Alternative way of authentication for internal usage only - should not be + exposed to public api. + "created_at": "2020-02-20 00:00:00", # Optional. Alternative + way of authentication for internal usage only - should not be exposed to + public api. + "created_by_user_email": "str", # Optional. Alternative way + of authentication for internal usage only - should not be exposed to + public api. + "created_by_user_id": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "dataset": { + "created_at": "2020-02-20 00:00:00", # Optional. + Time created at. + "dataset_name": "str", # Optional. Name of the + dataset. + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # + Optional. Default value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known + values are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". + "dataset_uuid": "str", # Optional. UUID of the + dataset. + "file_size": "str", # Optional. The size of the + dataset uploaded file in bytes. + "has_ground_truth": bool, # Optional. Does the + dataset have a ground truth column?. + "row_count": 0 # Optional. Number of rows in the + dataset. + }, + "dataset_name": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "dataset_uuid": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "description": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "latest_version_number_of_runs": 0, # Optional. Alternative + way of authentication for internal usage only - should not be exposed to + public api. + "metrics": [ { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. + "category": "METRIC_CATEGORY_UNSPECIFIED", # + Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known + values are: "METRIC_CATEGORY_UNSPECIFIED", + "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", + "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and + "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether a + metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the + metric is inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "metric_name": "str", # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "metric_rank": 0, # Optional. Alternative + way of authentication for internal usage only - should not be + exposed to public api. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known + values are: "METRIC_TYPE_UNSPECIFIED", + "METRIC_TYPE_GENERAL_QUALITY", "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum + value for the metric. + "range_min": 0.0 # Optional. The minimum + value for the metric. } ], - "data_source_uuids": [ - "str" # Optional. The indexing jobs. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - } - ], - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "name": "str", # Optional. Alternative way of authentication + for internal usage only - should not be exposed to public api. + "star_metric": { + "metric_uuid": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to + public api. + "name": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to + public api. + "success_threshold": 0.0, # Optional. The success + threshold for the star metric. This is a value that the metric must + reach to be considered successful. + "success_threshold_pct": 0 # Optional. The success + threshold for the star metric. This is a percentage value between 0 + and 100. + }, + "test_case_uuid": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "total_runs": 0, # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "updated_at": "2020-02-20 00:00:00", # Optional. Alternative + way of authentication for internal usage only - should not be exposed to + public api. + "updated_by_user_email": "str", # Optional. Alternative way + of authentication for internal usage only - should not be exposed to + public api. + "updated_by_user_id": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "version": 0 # Optional. Alternative way of authentication + for internal usage only - should not be exposed to public api. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. - "total": 0 # Optional. Total amount of items over all pages. - } + ] } # response body for status code(s): 404 response == { @@ -231974,9 +242183,7 @@ async def list_indexing_jobs( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_indexing_jobs_request( - page=page, - per_page=per_page, + _request = build_genai_list_evaluation_test_cases_request( headers=_headers, params=_params, ) @@ -232036,7 +242243,7 @@ async def list_indexing_jobs( return cast(JSON, deserialized) # type: ignore @overload - async def create_indexing_job( + async def create_evaluation_test_case( self, body: Optional[JSON] = None, *, @@ -232044,10 +242251,9 @@ async def create_indexing_job( **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Start Indexing Job for a Knowledge Base. + """Create Evaluation Test Case. - To start an indexing job for a knowledge base, send a POST request to - ``/v2/gen-ai/indexing_jobs``. + To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. :param body: Default value is None. :type body: JSON @@ -232063,86 +242269,29 @@ async def create_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "data_source_uuids": [ - "str" # Optional. List of data source ids to index, if none are - provided, all data sources will be indexed. + "agent_workspace_name": "str", # Optional. + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": [ + "str" # Optional. Full metric list to use for evaluation test case. ], - "knowledge_base_uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "workspace_uuid": "str" # Optional. The workspace uuid. } # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. - } + "test_case_uuid": "str" # Optional. Test"u2011case UUID. } # response body for status code(s): 404 response == { @@ -232158,7 +242307,7 @@ async def create_indexing_job( """ @overload - async def create_indexing_job( + async def create_evaluation_test_case( self, body: Optional[IO[bytes]] = None, *, @@ -232166,10 +242315,9 @@ async def create_indexing_job( **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Start Indexing Job for a Knowledge Base. + """Create Evaluation Test Case. - To start an indexing job for a knowledge base, send a POST request to - ``/v2/gen-ai/indexing_jobs``. + To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. :param body: Default value is None. :type body: IO[bytes] @@ -232185,77 +242333,7 @@ async def create_indexing_job( # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. - } + "test_case_uuid": "str" # Optional. Test"u2011case UUID. } # response body for status code(s): 404 response == { @@ -232271,14 +242349,13 @@ async def create_indexing_job( """ @distributed_trace_async - async def create_indexing_job( + async def create_evaluation_test_case( self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Start Indexing Job for a Knowledge Base. + """Create Evaluation Test Case. - To start an indexing job for a knowledge base, send a POST request to - ``/v2/gen-ai/indexing_jobs``. + To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] @@ -232291,86 +242368,29 @@ async def create_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "data_source_uuids": [ - "str" # Optional. List of data source ids to index, if none are - provided, all data sources will be indexed. + "agent_workspace_name": "str", # Optional. + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": [ + "str" # Optional. Full metric list to use for evaluation test case. ], - "knowledge_base_uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "workspace_uuid": "str" # Optional. The workspace uuid. } # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. - } + "test_case_uuid": "str" # Optional. Test"u2011case UUID. } # response body for status code(s): 404 response == { @@ -232416,7 +242436,7 @@ async def create_indexing_job( else: _json = None - _request = build_genai_create_indexing_job_request( + _request = build_genai_create_evaluation_test_case_request( content_type=content_type, json=_json, content=_content, @@ -232479,17 +242499,23 @@ async def create_indexing_job( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def list_indexing_job_data_sources( - self, indexing_job_uuid: str, **kwargs: Any + async def list_evaluation_runs_by_test_case( + self, + evaluation_test_case_uuid: str, + *, + evaluation_test_case_version: Optional[int] = None, + **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """List Data Sources for Indexing Job for a Knowledge Base. + """List Evaluation Runs by Test Case. - To list all datasources for an indexing job, send a GET request to - ``/v2/gen-ai/indexing_jobs/{indexing_job_uuid}/data_sources``. + To list all evaluation runs by test case, send a GET request to + ``/v2/gen-ai/evaluation_test_cases/{evaluation_test_case_uuid}/evaluation_runs``. - :param indexing_job_uuid: Uuid of the indexing job. Required. - :type indexing_job_uuid: str + :param evaluation_test_case_uuid: Evaluation run UUID. Required. + :type evaluation_test_case_uuid: str + :keyword evaluation_test_case_version: Version of the test case. Default value is None. + :paramtype evaluation_test_case_version: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -232499,40 +242525,82 @@ async def list_indexing_job_data_sources( # response body for status code(s): 200 response == { - "indexed_data_sources": [ + "evaluation_runs": [ { - "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp - when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the indexed - data source. - "error_details": "str", # Optional. A detailed error + "agent_deleted": bool, # Optional. Whether agent is deleted. + "agent_deployment_name": "str", # Optional. The agent + deployment name. + "agent_name": "str", # Optional. Agent name. + "agent_uuid": "str", # Optional. Agent UUID. + "agent_version_hash": "str", # Optional. Version hash. + "agent_workspace_uuid": "str", # Optional. Agent workspace + uuid. + "created_by_user_email": "str", # Optional. List of + evaluation runs. + "created_by_user_id": "str", # Optional. List of evaluation + runs. + "error_description": "str", # Optional. The error description. - "error_msg": "str", # Optional. A string code provinding a - hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count of files - that have failed. - "indexed_file_count": "str", # Optional. Total count of - files that have been indexed. - "indexed_item_count": "str", # Optional. Total count of - files that have been indexed. - "removed_item_count": "str", # Optional. Total count of - files that have been removed. - "skipped_item_count": "str", # Optional. Total count of - files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. Timestamp - when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default - value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of files in - data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size of - files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file count in - the data source. + "evaluation_run_uuid": "str", # Optional. Evaluation run + UUID. + "evaluation_test_case_workspace_uuid": "str", # Optional. + Evaluation test case workspace uuid. + "finished_at": "2020-02-20 00:00:00", # Optional. Run end + time. + "pass_status": bool, # Optional. The pass status of the + evaluation run based on the star metric. + "queued_at": "2020-02-20 00:00:00", # Optional. Run queued + time. + "run_level_metric_results": [ + { + "error_description": "str", # Optional. + Error description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric + name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value + of the metric as a number. + "reasoning": "str", # Optional. Reasoning of + the metric result. + "string_value": "str" # Optional. The value + of the metric as a string. + } + ], + "run_name": "str", # Optional. Run name. + "star_metric_result": { + "error_description": "str", # Optional. Error + description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the + metric as a number. + "reasoning": "str", # Optional. Reasoning of the + metric result. + "string_value": "str" # Optional. The value of the + metric as a string. + }, + "started_at": "2020-02-20 00:00:00", # Optional. Run start + time. + "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. + Default value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run + Statuses. Known values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", + "EVALUATION_RUN_QUEUED", "EVALUATION_RUN_RUNNING_DATASET", + "EVALUATION_RUN_EVALUATING_RESULTS", "EVALUATION_RUN_CANCELLING", + "EVALUATION_RUN_CANCELLED", "EVALUATION_RUN_SUCCESSFUL", + "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and "EVALUATION_RUN_FAILED". + "test_case_description": "str", # Optional. Test case + description. + "test_case_name": "str", # Optional. Test case name. + "test_case_uuid": "str", # Optional. Test-case UUID. + "test_case_version": 0 # Optional. Test-case-version. } ] } @@ -232566,8 +242634,9 @@ async def list_indexing_job_data_sources( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_indexing_job_data_sources_request( - indexing_job_uuid=indexing_job_uuid, + _request = build_genai_list_evaluation_runs_by_test_case_request( + evaluation_test_case_uuid=evaluation_test_case_uuid, + evaluation_test_case_version=evaluation_test_case_version, headers=_headers, params=_params, ) @@ -232627,17 +242696,23 @@ async def list_indexing_job_data_sources( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def get_indexing_job_details_signed_url( - self, indexing_job_uuid: str, **kwargs: Any + async def get_evaluation_test_case( + self, + test_case_uuid: str, + *, + evaluation_test_case_version: Optional[int] = None, + **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Get Signed URL for Indexing Job Details. + """Retrieve Information About an Existing Evaluation Test Case. - To get a signed URL for indexing job details, send a GET request to - ``/v2/gen-ai/indexing_jobs/{uuid}/details_signed_url``. + To retrive information about an existing evaluation test case, send a GET request to + ``/v2/gen-ai/evaluation_test_case/{test_case_uuid}``. - :param indexing_job_uuid: The uuid of the indexing job. Required. - :type indexing_job_uuid: str + :param test_case_uuid: The test case uuid to retrieve. Required. + :type test_case_uuid: str + :keyword evaluation_test_case_version: Version of the test case. Default value is None. + :paramtype evaluation_test_case_version: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -232647,8 +242722,86 @@ async def get_indexing_job_details_signed_url( # response body for status code(s): 200 response == { - "signed_url": "str" # Optional. The signed url for downloading the indexing - job details. + "evaluation_test_case": { + "archived_at": "2020-02-20 00:00:00", # Optional. + "created_at": "2020-02-20 00:00:00", # Optional. + "created_by_user_email": "str", # Optional. + "created_by_user_id": "str", # Optional. + "dataset": { + "created_at": "2020-02-20 00:00:00", # Optional. Time + created at. + "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # + Optional. Default value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known + values are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". + "dataset_uuid": "str", # Optional. UUID of the dataset. + "file_size": "str", # Optional. The size of the dataset + uploaded file in bytes. + "has_ground_truth": bool, # Optional. Does the dataset have + a ground truth column?. + "row_count": 0 # Optional. Number of rows in the dataset. + }, + "dataset_name": "str", # Optional. + "dataset_uuid": "str", # Optional. + "description": "str", # Optional. + "latest_version_number_of_runs": 0, # Optional. + "metrics": [ + { + "category": "METRIC_CATEGORY_UNSPECIFIED", # + Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known + values are: "METRIC_CATEGORY_UNSPECIFIED", + "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", + "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", + # Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope + that determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to + agent metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. + "metric_name": "str", # Optional. + "metric_rank": 0, # Optional. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values + are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for + the metric. + "range_min": 0.0 # Optional. The minimum value for + the metric. + } + ], + "name": "str", # Optional. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold + for the star metric. This is a value that the metric must reach to be + considered successful. + "success_threshold_pct": 0 # Optional. The success threshold + for the star metric. This is a percentage value between 0 and 100. + }, + "test_case_uuid": "str", # Optional. + "total_runs": 0, # Optional. + "updated_at": "2020-02-20 00:00:00", # Optional. + "updated_by_user_email": "str", # Optional. + "updated_by_user_id": "str", # Optional. + "version": 0 # Optional. + } } # response body for status code(s): 404 response == { @@ -232680,8 +242833,9 @@ async def get_indexing_job_details_signed_url( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_indexing_job_details_signed_url_request( - indexing_job_uuid=indexing_job_uuid, + _request = build_genai_get_evaluation_test_case_request( + test_case_uuid=test_case_uuid, + evaluation_test_case_version=evaluation_test_case_version, headers=_headers, params=_params, ) @@ -232740,16 +242894,140 @@ async def get_indexing_job_details_signed_url( return cast(JSON, deserialized) # type: ignore + @overload + async def update_evaluation_test_case( + self, + test_case_uuid: str, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Update an Evaluation Test Case. + + To update an evaluation test-case send a PUT request to + ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. + + :param test_case_uuid: Test-case UUID to update. Required. + :type test_case_uuid: str + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": { + "metric_uuids": [ + "str" # Optional. + ] + }, + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "test_case_uuid": "str" # Optional. Test-case UUID to update. + } + + # response body for status code(s): 200 + response == { + "test_case_uuid": "str", # Optional. + "version": 0 # Optional. The new verson of the test case. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + async def update_evaluation_test_case( + self, + test_case_uuid: str, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Update an Evaluation Test Case. + + To update an evaluation test-case send a PUT request to + ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. + + :param test_case_uuid: Test-case UUID to update. Required. + :type test_case_uuid: str + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "test_case_uuid": "str", # Optional. + "version": 0 # Optional. The new verson of the test case. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + @distributed_trace_async - async def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: + async def update_evaluation_test_case( + self, + test_case_uuid: str, + body: Optional[Union[JSON, IO[bytes]]] = None, + **kwargs: Any + ) -> JSON: # pylint: disable=line-too-long - """Retrieve Status of Indexing Job for a Knowledge Base. + """Update an Evaluation Test Case. - To get status of an indexing Job for a knowledge base, send a GET request to - ``/v2/gen-ai/indexing_jobs/{uuid}``. + To update an evaluation test-case send a PUT request to + ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. - :param uuid: Indexing job id. Required. - :type uuid: str + :param test_case_uuid: Test-case UUID to update. Required. + :type test_case_uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -232757,78 +243035,258 @@ async def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": { + "metric_uuids": [ + "str" # Optional. + ] + }, + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "test_case_uuid": "str" # Optional. Test-case UUID to update. + } + # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. + "test_case_uuid": "str", # Optional. + "version": 0 # Optional. The new verson of the test case. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + cls: ClsType[JSON] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_update_evaluation_test_case_request( + test_case_uuid=test_case_uuid, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def list_indexing_jobs( + self, + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """List Indexing Jobs for a Knowledge Base. + + To list all indexing jobs for a knowledge base, send a GET request to + ``/v2/gen-ai/indexing_jobs``. + + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "jobs": [ + { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. The indexing jobs. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + } + ], + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. } } # response body for status code(s): 404 @@ -232861,8 +243319,9 @@ async def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_indexing_job_request( - uuid=uuid, + _request = build_genai_list_indexing_jobs_request( + page=page, + per_page=per_page, headers=_headers, params=_params, ) @@ -232922,22 +243381,19 @@ async def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - async def cancel_indexing_job( + async def create_indexing_job( self, - uuid: str, body: Optional[JSON] = None, *, content_type: str = "application/json", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Cancel Indexing Job for a Knowledge Base. + """Start Indexing Job for a Knowledge Base. - To cancel an indexing job for a knowledge base, send a PUT request to - ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + To start an indexing job for a knowledge base, send a POST request to + ``/v2/gen-ai/indexing_jobs``. - :param uuid: A unique identifier for an indexing job. Required. - :type uuid: str :param body: Default value is None. :type body: JSON :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. @@ -232952,7 +243408,11 @@ async def cancel_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "uuid": "str" # Optional. A unique identifier for an indexing job. + "data_source_uuids": [ + "str" # Optional. List of data source ids to index, if none are + provided, all data sources will be indexed. + ], + "knowledge_base_uuid": "str" # Optional. Knowledge base id. } # response body for status code(s): 200 @@ -233043,22 +243503,19 @@ async def cancel_indexing_job( """ @overload - async def cancel_indexing_job( + async def create_indexing_job( self, - uuid: str, body: Optional[IO[bytes]] = None, *, content_type: str = "application/json", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Cancel Indexing Job for a Knowledge Base. + """Start Indexing Job for a Knowledge Base. - To cancel an indexing job for a knowledge base, send a PUT request to - ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + To start an indexing job for a knowledge base, send a POST request to + ``/v2/gen-ai/indexing_jobs``. - :param uuid: A unique identifier for an indexing job. Required. - :type uuid: str :param body: Default value is None. :type body: IO[bytes] :keyword content_type: Body Parameter content-type. Content type parameter for binary body. @@ -233159,17 +243616,15 @@ async def cancel_indexing_job( """ @distributed_trace_async - async def cancel_indexing_job( - self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + async def create_indexing_job( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Cancel Indexing Job for a Knowledge Base. + """Start Indexing Job for a Knowledge Base. - To cancel an indexing job for a knowledge base, send a PUT request to - ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + To start an indexing job for a knowledge base, send a POST request to + ``/v2/gen-ai/indexing_jobs``. - :param uuid: A unique identifier for an indexing job. Required. - :type uuid: str :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] :return: JSON object @@ -233181,7 +243636,11 @@ async def cancel_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "uuid": "str" # Optional. A unique identifier for an indexing job. + "data_source_uuids": [ + "str" # Optional. List of data source ids to index, if none are + provided, all data sources will be indexed. + ], + "knowledge_base_uuid": "str" # Optional. Knowledge base id. } # response body for status code(s): 200 @@ -233302,8 +243761,7 @@ async def cancel_indexing_job( else: _json = None - _request = build_genai_cancel_indexing_job_request( - uuid=uuid, + _request = build_genai_create_indexing_job_request( content_type=content_type, json=_json, content=_content, @@ -233366,22 +243824,17 @@ async def cancel_indexing_job( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def list_knowledge_bases( - self, - *, - page: Optional[int] = None, - per_page: Optional[int] = None, - **kwargs: Any + async def list_indexing_job_data_sources( + self, indexing_job_uuid: str, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """List Knowledge Bases. + """List Data Sources for Indexing Job for a Knowledge Base. - To list all knowledge bases, send a GET request to ``/v2/gen-ai/knowledge_bases``. + To list all datasources for an indexing job, send a GET request to + ``/v2/gen-ai/indexing_jobs/{indexing_job_uuid}/data_sources``. - :keyword page: Page number. Default value is None. - :paramtype page: int - :keyword per_page: Items per page. Default value is None. - :paramtype per_page: int + :param indexing_job_uuid: Uuid of the indexing job. Required. + :type indexing_job_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -233391,129 +243844,1021 @@ async def list_knowledge_bases( # response body for status code(s): 200 response == { - "knowledge_bases": [ + "indexed_data_sources": [ { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time - when the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "database_id": "str", # Optional. The knowledge bases. - "embedding_model_uuid": "str", # Optional. The knowledge - bases. - "is_public": bool, # Optional. Whether the knowledge base is - public or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. - Creation date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 - 00:00:00", # Optional. Timestamp when data source completed - indexing. - "data_source_uuid": "str", # - Optional. Uuid of the indexed data source. - "error_details": "str", # Optional. - A detailed error description. - "error_msg": "str", # Optional. A - string code provinding a hint which part of the system - experienced an error. - "failed_item_count": "str", # - Optional. Total count of files that have failed. - "indexed_file_count": "str", # - Optional. Total count of files that have been indexed. - "indexed_item_count": "str", # - Optional. Total count of files that have been indexed. - "removed_item_count": "str", # - Optional. Total count of files that have been removed. - "skipped_item_count": "str", # - Optional. Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", - # Optional. Timestamp when data source started indexing. - "status": - "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default value is - "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", - "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. - Total size of files in data source in bytes. - "total_bytes_indexed": "str", # - Optional. Total size of files in data source in bytes that - have been indexed. - "total_file_count": "str" # - Optional. Total file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean - value to determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge - base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. - Default value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. - Default value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This - field is deprecated]. - "total_datasources": 0, # Optional. Number of - datasources being indexed. - "total_tokens": "str", # Optional. Total Tokens - Consumed By the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. - Last modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. The knowledge bases. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is - enabled for retrieval. - "model": "str" # Optional. Reranker model internal - name. - }, - "tags": [ - "str" # Optional. Tags to organize related - resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "user_id": "str", # Optional. Id of user that created the - knowledge base. - "uuid": "str" # Optional. Unique id for knowledge base. - } - ], - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp + when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the indexed + data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code provinding a + hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count of files + that have failed. + "indexed_file_count": "str", # Optional. Total count of + files that have been indexed. + "indexed_item_count": "str", # Optional. Total count of + files that have been indexed. + "removed_item_count": "str", # Optional. Total count of + files that have been removed. + "skipped_item_count": "str", # Optional. Total count of + files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp + when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default + value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of files in + data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size of + files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file count in + the data source. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_list_indexing_job_data_sources_request( + indexing_job_uuid=indexing_job_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def get_indexing_job_details_signed_url( + self, indexing_job_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Get Signed URL for Indexing Job Details. + + To get a signed URL for indexing job details, send a GET request to + ``/v2/gen-ai/indexing_jobs/{uuid}/details_signed_url``. + + :param indexing_job_uuid: The uuid of the indexing job. Required. + :type indexing_job_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "signed_url": "str" # Optional. The signed url for downloading the indexing + job details. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_indexing_job_details_signed_url_request( + indexing_job_uuid=indexing_job_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Retrieve Status of Indexing Job for a Knowledge Base. + + To get status of an indexing Job for a knowledge base, send a GET request to + ``/v2/gen-ai/indexing_jobs/{uuid}``. + + :param uuid: Indexing job id. Required. + :type uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_indexing_job_request( + uuid=uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @overload + async def cancel_indexing_job( + self, + uuid: str, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Cancel Indexing Job for a Knowledge Base. + + To cancel an indexing job for a knowledge base, send a PUT request to + ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + + :param uuid: A unique identifier for an indexing job. Required. + :type uuid: str + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "uuid": "str" # Optional. A unique identifier for an indexing job. + } + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + async def cancel_indexing_job( + self, + uuid: str, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Cancel Indexing Job for a Knowledge Base. + + To cancel an indexing job for a knowledge base, send a PUT request to + ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + + :param uuid: A unique identifier for an indexing job. Required. + :type uuid: str + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace_async + async def cancel_indexing_job( + self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Cancel Indexing Job for a Knowledge Base. + + To cancel an indexing job for a knowledge base, send a PUT request to + ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + + :param uuid: A unique identifier for an indexing job. Required. + :type uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "uuid": "str" # Optional. A unique identifier for an indexing job. + } + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + cls: ClsType[JSON] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_cancel_indexing_job_request( + uuid=uuid, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def list_knowledge_bases( + self, + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """List Knowledge Bases. + + To list all knowledge bases, send a GET request to ``/v2/gen-ai/knowledge_bases``. + + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "knowledge_bases": [ + { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time + when the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "database_id": "str", # Optional. The knowledge bases. + "embedding_model_uuid": "str", # Optional. The knowledge + bases. + "is_public": bool, # Optional. Whether the knowledge base is + public or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 + 00:00:00", # Optional. Timestamp when data source completed + indexing. + "data_source_uuid": "str", # + Optional. Uuid of the indexed data source. + "error_details": "str", # Optional. + A detailed error description. + "error_msg": "str", # Optional. A + string code provinding a hint which part of the system + experienced an error. + "failed_item_count": "str", # + Optional. Total count of files that have failed. + "indexed_file_count": "str", # + Optional. Total count of files that have been indexed. + "indexed_item_count": "str", # + Optional. Total count of files that have been indexed. + "removed_item_count": "str", # + Optional. Total count of files that have been removed. + "skipped_item_count": "str", # + Optional. Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", + # Optional. Timestamp when data source started indexing. + "status": + "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default value is + "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", + "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. + Total size of files in data source in bytes. + "total_bytes_indexed": "str", # + Optional. Total size of files in data source in bytes that + have been indexed. + "total_file_count": "str" # + Optional. Total file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean + value to determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge + base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. + Default value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. + Default value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This + field is deprecated]. + "total_datasources": 0, # Optional. Number of + datasources being indexed. + "total_tokens": "str", # Optional. Total Tokens + Consumed By the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. The knowledge bases. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is + enabled for retrieval. + "model": "str" # Optional. Reranker model internal + name. + }, + "tags": [ + "str" # Optional. Tags to organize related + resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "user_id": "str", # Optional. Id of user that created the + knowledge base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + ], + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. "total": 0 # Optional. Total amount of items over all pages. } } @@ -233652,18 +244997,21 @@ async def create_knowledge_base( data_source_details. "bucket_region": "str", # Optional. Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The - chunking algorithm to use for processing data sources. **Note: This - feature requires enabling the knowledgebase enhancements feature preview - flag.**. Known values are: "CHUNKING_ALGORITHM_UNKNOWN", - "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", - "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # + Optional. Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values + are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and + "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical - options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Optional data + sources to attach at creation. Omit or use an empty list to create + the knowledge base without sources, then add sources (with chunking + strategy and sizes) using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic @@ -233690,10 +245038,14 @@ async def create_knowledge_base( you can obrain a refresh token by following the oauth2 flow. see /v2/gen-ai/oauth2/google/tokens for reference. }, - "item_path": "str", # Optional. The data sources to use for - this knowledge base. See `Organize Data Sources - `_ - for more information on data sources best practices. + "item_path": "str", # Optional. Optional data sources to + attach at creation. Omit or use an empty list to create the knowledge + base without sources, then add sources (with chunking strategy and sizes) + using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. "spaces_data_source": { "bucket_name": "str", # Optional. Spaces bucket name. @@ -233735,6 +245087,11 @@ async def create_knowledge_base( retrieval. "model": "str" # Optional. Reranker model internal name. }, + "size": "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", # Optional. Default value is + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED". Known values are: + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", "OPEN_SEARCH_PLAN_SIZE_SMALL", + "OPEN_SEARCH_PLAN_SIZE_MEDIUM", "OPEN_SEARCH_PLAN_SIZE_LARGE", and + "OPEN_SEARCH_PLAN_SIZE_EXTRA_LARGE". "tags": [ "str" # Optional. Tags to organize your knowledge base. ], @@ -234040,18 +245397,21 @@ async def create_knowledge_base( data_source_details. "bucket_region": "str", # Optional. Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The - chunking algorithm to use for processing data sources. **Note: This - feature requires enabling the knowledgebase enhancements feature preview - flag.**. Known values are: "CHUNKING_ALGORITHM_UNKNOWN", - "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", - "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # + Optional. Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values + are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and + "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical - options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Optional data + sources to attach at creation. Omit or use an empty list to create + the knowledge base without sources, then add sources (with chunking + strategy and sizes) using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic @@ -234078,10 +245438,14 @@ async def create_knowledge_base( you can obrain a refresh token by following the oauth2 flow. see /v2/gen-ai/oauth2/google/tokens for reference. }, - "item_path": "str", # Optional. The data sources to use for - this knowledge base. See `Organize Data Sources - `_ - for more information on data sources best practices. + "item_path": "str", # Optional. Optional data sources to + attach at creation. Omit or use an empty list to create the knowledge + base without sources, then add sources (with chunking strategy and sizes) + using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. "spaces_data_source": { "bucket_name": "str", # Optional. Spaces bucket name. @@ -234123,6 +245487,11 @@ async def create_knowledge_base( retrieval. "model": "str" # Optional. Reranker model internal name. }, + "size": "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", # Optional. Default value is + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED". Known values are: + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", "OPEN_SEARCH_PLAN_SIZE_SMALL", + "OPEN_SEARCH_PLAN_SIZE_MEDIUM", "OPEN_SEARCH_PLAN_SIZE_LARGE", and + "OPEN_SEARCH_PLAN_SIZE_EXTRA_LARGE". "tags": [ "str" # Optional. Tags to organize your knowledge base. ], @@ -234658,18 +246027,14 @@ async def list_knowledge_base_data_sources( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The - chunking algorithm to use for processing data sources. **Note: This - feature requires enabling the knowledgebase enhancements feature preview - flag.**. Known values are: "CHUNKING_ALGORITHM_UNKNOWN", - "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", - "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # + Optional. Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values + are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and + "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical - options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. The data sources. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic @@ -234908,17 +246273,14 @@ async def create_knowledge_base_data_source( "region": "str", # Optional. Region of bucket. "secret_key": "str" # Optional. The AWS Secret Key. }, - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -234958,17 +246320,15 @@ async def create_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235108,17 +246468,15 @@ async def create_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235251,17 +246609,14 @@ async def create_knowledge_base_data_source( "region": "str", # Optional. Region of bucket. "secret_key": "str" # Optional. The AWS Secret Key. }, - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235301,17 +246656,15 @@ async def create_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235539,17 +246892,14 @@ async def update_knowledge_base_data_source( # JSON input template you can fill out and use as your body input. body = { - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235568,17 +246918,15 @@ async def update_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235721,17 +247069,15 @@ async def update_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235860,17 +247206,14 @@ async def update_knowledge_base_data_source( # JSON input template you can fill out and use as your body input. body = { - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -235889,17 +247232,15 @@ async def update_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -236095,19 +247436,1346 @@ async def update_knowledge_base_data_source( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def delete_knowledge_base_data_source( - self, knowledge_base_uuid: str, data_source_uuid: str, **kwargs: Any + async def delete_knowledge_base_data_source( + self, knowledge_base_uuid: str, data_source_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Delete a Data Source from a Knowledge Base. + + To delete a data source from a knowledge base, send a DELETE request to + ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/data_sources/{data_source_uuid}``. + + :param knowledge_base_uuid: Knowledge base id. Required. + :type knowledge_base_uuid: str + :param data_source_uuid: Data source id. Required. + :type data_source_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "data_source_uuid": "str", # Optional. Data source id. + "knowledge_base_uuid": "str" # Optional. Knowledge base id. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_delete_knowledge_base_data_source_request( + knowledge_base_uuid=knowledge_base_uuid, + data_source_uuid=data_source_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def list_indexing_jobs_by_knowledge_base( + self, knowledge_base_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """List Indexing Jobs for a Knowledge Base. + + To list latest 15 indexing jobs for a knowledge base, send a GET request to + ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/indexing_jobs``. + + :param knowledge_base_uuid: Knowledge base uuid in string. Required. + :type knowledge_base_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "jobs": [ + { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. The indexing jobs. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + } + ], + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_list_indexing_jobs_by_knowledge_base_request( + knowledge_base_uuid=knowledge_base_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Retrieve Information About an Existing Knowledge Base. + + To retrive information about an existing knowledge base, send a GET request to + ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "database_status": "CREATING", # Optional. Default value is "CREATING". + Known values are: "CREATING", "ONLINE", "POWEROFF", "REBUILDING", "REBALANCING", + "DECOMMISSIONED", "FORKING", "MIGRATING", "RESIZING", "RESTORING", "POWERING_ON", + "UNHEALTHY", and "UPGRADING". + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_knowledge_base_request( + uuid=uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @overload + async def update_knowledge_base( + self, + uuid: str, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Update a Knowledge Base. + + To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "database_id": "str", # Optional. The id of the DigitalOcean database this + knowledge base will use, optional. + "name": "str", # Optional. Knowledge base name. + "project_id": "str", # Optional. The id of the DigitalOcean project this + knowledge base will belong to. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled for + retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize your knowledge base. + ], + "uuid": "str" # Optional. Knowledge base id. + } + + # response body for status code(s): 200 + response == { + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + async def update_knowledge_base( + self, + uuid: str, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Update a Knowledge Base. + + To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace_async + async def update_knowledge_base( + self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Update a Knowledge Base. + + To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "database_id": "str", # Optional. The id of the DigitalOcean database this + knowledge base will use, optional. + "name": "str", # Optional. Knowledge base name. + "project_id": "str", # Optional. The id of the DigitalOcean project this + knowledge base will belong to. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled for + retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize your knowledge base. + ], + "uuid": "str" # Optional. Knowledge base id. + } + + # response body for status code(s): 200 + response == { + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + cls: ClsType[JSON] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_update_knowledge_base_request( + uuid=uuid, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Delete a Knowledge Base. + + To delete a knowledge base, send a DELETE request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "uuid": "str" # Optional. The id of the deleted knowledge base. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_delete_knowledge_base_request( + uuid=uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @overload + async def create_model_eval_dataset_upload_presigned_urls( # pylint: disable=name-too-long + self, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Model Evaluation Dataset File Upload. + + To create presigned URLs for model evaluation dataset file upload, send a POST request to + ``/v2/genai/model_evaluation/datasets/file_upload_presigned_urls``. + + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + async def create_model_eval_dataset_upload_presigned_urls( # pylint: disable=name-too-long + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Model Evaluation Dataset File Upload. + + To create presigned URLs for model evaluation dataset file upload, send a POST request to + ``/v2/genai/model_evaluation/datasets/file_upload_presigned_urls``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace_async + async def create_model_eval_dataset_upload_presigned_urls( # pylint: disable=name-too-long + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Delete a Data Source from a Knowledge Base. + """Create Presigned URLs for Model Evaluation Dataset File Upload. - To delete a data source from a knowledge base, send a DELETE request to - ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/data_sources/{data_source_uuid}``. + To create presigned URLs for model evaluation dataset file upload, send a POST request to + ``/v2/genai/model_evaluation/datasets/file_upload_presigned_urls``. - :param knowledge_base_uuid: Knowledge base id. Required. - :type knowledge_base_uuid: str - :param data_source_uuid: Data source id. Required. - :type data_source_uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -236115,10 +248783,33 @@ async def delete_knowledge_base_data_source( Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + # response body for status code(s): 200 response == { - "data_source_uuid": "str", # Optional. Data source id. - "knowledge_base_uuid": "str" # Optional. Knowledge base id. + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] } # response body for status code(s): 404 response == { @@ -236145,14 +248836,29 @@ async def delete_knowledge_base_data_source( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = kwargs.pop("headers", {}) or {} + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_delete_knowledge_base_data_source_request( - knowledge_base_uuid=knowledge_base_uuid, - data_source_uuid=data_source_uuid, + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_create_model_eval_dataset_upload_presigned_urls_request( + content_type=content_type, + json=_json, + content=_content, headers=_headers, params=_params, ) @@ -236212,17 +248918,13 @@ async def delete_knowledge_base_data_source( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def list_indexing_jobs_by_knowledge_base( - self, knowledge_base_uuid: str, **kwargs: Any - ) -> JSON: + async def list_model_evaluation_metrics(self, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Indexing Jobs for a Knowledge Base. + """List Model Evaluation Metrics. - To list latest 15 indexing jobs for a knowledge base, send a GET request to - ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/indexing_jobs``. + To list all available metrics for model evaluation, send a GET request to + ``/v2/genai/model_evaluation_metrics``. - :param knowledge_base_uuid: Knowledge base uuid in string. Required. - :type knowledge_base_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -236232,97 +248934,47 @@ async def list_indexing_jobs_by_knowledge_base( # response body for status code(s): 200 response == { - "jobs": [ + "metrics": [ { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. The indexing jobs. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - } - ], - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "category": "METRIC_CATEGORY_UNSPECIFIED", # Optional. + Default value is "METRIC_CATEGORY_UNSPECIFIED". Known values are: + "METRIC_CATEGORY_UNSPECIFIED", "METRIC_CATEGORY_CORRECTNESS", + "METRIC_CATEGORY_USER_OUTCOMES", "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. List of model evaluation + metrics. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", # + Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. List of model evaluation + metrics. + "metric_name": "str", # Optional. List of model evaluation + metrics. + "metric_rank": 0, # Optional. List of model evaluation + metrics. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. + Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. List of model evaluation + metrics. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for the + metric. + "range_min": 0.0 # Optional. The minimum value for the + metric. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. - "total": 0 # Optional. Total amount of items over all pages. - } + ] } # response body for status code(s): 404 response == { @@ -236354,8 +249006,7 @@ async def list_indexing_jobs_by_knowledge_base( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_indexing_jobs_by_knowledge_base_request( - knowledge_base_uuid=knowledge_base_uuid, + _request = build_genai_list_model_evaluation_metrics_request( headers=_headers, params=_params, ) @@ -236415,15 +249066,33 @@ async def list_indexing_jobs_by_knowledge_base( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + async def list_model_evaluation_runs( + self, + *, + eval_preset_uuid: Optional[str] = None, + status: str = "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any + ) -> JSON: # pylint: disable=line-too-long - """Retrieve Information About an Existing Knowledge Base. - - To retrive information about an existing knowledge base, send a GET request to - ``/v2/gen-ai/knowledge_bases/{uuid}``. - - :param uuid: Knowledge base id. Required. - :type uuid: str + """List Model Evaluation Runs. + + To list model evaluation runs, send a GET request to ``/v2/genai/model_evaluation_runs``. + + :keyword eval_preset_uuid: UUID of the evaluation preset to filter by. Default value is None. + :paramtype eval_preset_uuid: str + :keyword status: Filter by evaluation run status. Known values are: + "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", "MODEL_EVALUATION_RUN_QUEUED", + "MODEL_EVALUATION_RUN_RUNNING_DATASET", "MODEL_EVALUATION_RUN_EVALUATING_RESULTS", + "MODEL_EVALUATION_RUN_CANCELLING", "MODEL_EVALUATION_RUN_CANCELLED", + "MODEL_EVALUATION_RUN_SUCCESSFUL", "MODEL_EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "MODEL_EVALUATION_RUN_FAILED". Default value is "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED". + :paramtype status: str + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -236433,112 +249102,56 @@ async def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "database_status": "CREATING", # Optional. Default value is "CREATING". - Known values are: "CREATING", "ONLINE", "POWEROFF", "REBUILDING", "REBALANCING", - "DECOMMISSIONED", "FORKING", "MIGRATING", "RESIZING", "RESTORING", "POWERING_ON", - and "UNHEALTHY". - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "runs": [ + { + "candidate_model_name": "str", # Optional. Name of the + candidate model being evaluated. + "candidate_model_source": + "CANDIDATE_MODEL_SOURCE_SERVERLESS", # Optional. Default value is + "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether inference runs against the + serverless platform, a dedicated deployment, or a model router. Known + values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. UUID of the + candidate model being evaluated. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp + when the run was created. + "dataset_name": "str", # Optional. Name of the dataset used + for evaluation. + "dataset_uuid": "str", # Optional. UUID of the dataset used + for evaluation. + "eval_run_uuid": "str", # Optional. UUID of the evaluation + run. + "judge_model_name": "str", # Optional. Summary view of + evaluation runs for the run history list. + "judge_model_uuid": "str", # Optional. Judge model used to + score responses. + "name": "str", # Optional. Name of the evaluation run. + "status": "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED" # + Optional. Default value is "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED". + Model Evaluation Run Statuses. Known values are: + "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", "MODEL_EVALUATION_RUN_QUEUED", + "MODEL_EVALUATION_RUN_RUNNING_DATASET", + "MODEL_EVALUATION_RUN_EVALUATING_RESULTS", + "MODEL_EVALUATION_RUN_CANCELLING", "MODEL_EVALUATION_RUN_CANCELLED", + "MODEL_EVALUATION_RUN_SUCCESSFUL", + "MODEL_EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "MODEL_EVALUATION_RUN_FAILED". + } + ] } # response body for status code(s): 404 response == { @@ -236570,8 +249183,11 @@ async def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_knowledge_base_request( - uuid=uuid, + _request = build_genai_list_model_evaluation_runs_request( + eval_preset_uuid=eval_preset_uuid, + status=status, + page=page, + per_page=per_page, headers=_headers, params=_params, ) @@ -236631,21 +249247,18 @@ async def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - async def update_knowledge_base( + async def create_model_evaluation_run( self, - uuid: str, body: Optional[JSON] = None, *, content_type: str = "application/json", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Update a Knowledge Base. + """Create Model Evaluation Run. - To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To create a model evaluation run, send a POST request to ``/v2/genai/model_evaluation_runs``. - :param uuid: Knowledge base id. Required. - :type uuid: str :param body: Default value is None. :type body: JSON :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. @@ -236660,128 +249273,55 @@ async def update_knowledge_base( # JSON input template you can fill out and use as your body input. body = { - "database_id": "str", # Optional. The id of the DigitalOcean database this - knowledge base will use, optiona. - "embedding_model_uuid": "str", # Optional. Identifier for the foundation - model. - "name": "str", # Optional. Knowledge base name. - "project_id": "str", # Optional. The id of the DigitalOcean project this - knowledge base will belong to. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled for - retrieval. - "model": "str" # Optional. Reranker model internal name. + "candidate_inference_config": { + "max_tokens": 0, # Optional. Inference configuration for the + candidate model during evaluation. + "stop_token": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "system_prompt": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "temperature": 0.0 # Optional. Inference configuration for the + candidate model during evaluation. }, - "tags": [ - "str" # Optional. Tags to organize your knowledge base. + "candidate_model_name": "str", # Optional. Model slug used to call the + candidate model API. For dedicated inference, this is the model slug from the + deployment. For serverless, this should match the model's internal name. + "candidate_model_source": "CANDIDATE_MODEL_SOURCE_SERVERLESS", # Optional. + Default value is "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether inference runs + against the serverless platform, a dedicated deployment, or a model router. Known + values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. UUID of the candidate model to + evaluate. + "dataset_uuid": "str", # Optional. UUID of the dataset to use for + evaluation. + "eval_preset_uuid": "str", # Optional. + "judge_model_uuid": "str", # Optional. UUID of the judge model used to score + responses. + "metric_uuids": [ + "str" # Optional. UUIDs of metrics to evaluate (selected from + ListModelEvaluationMetrics). ], - "uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. + "preset_name": "str", # Optional. + "save_as_preset": bool, # Optional. .. role:: raw-html-m2r(raw) :format: + html If true, saves the inline config as a reusable preset"" + :raw-html-m2r:`
` Ignored when eval_preset_uuid is provided. + "source": "str", # Optional. Source of the run creation (api, sdk, cli). + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + } } # response body for status code(s): 200 response == { - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "eval_run_uuid": "str" # Optional. UUID of the created evaluation run. } # response body for status code(s): 404 response == { @@ -236797,21 +249337,18 @@ async def update_knowledge_base( """ @overload - async def update_knowledge_base( + async def create_model_evaluation_run( self, - uuid: str, body: Optional[IO[bytes]] = None, *, content_type: str = "application/json", **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Update a Knowledge Base. + """Create Model Evaluation Run. - To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To create a model evaluation run, send a POST request to ``/v2/genai/model_evaluation_runs``. - :param uuid: Knowledge base id. Required. - :type uuid: str :param body: Default value is None. :type body: IO[bytes] :keyword content_type: Body Parameter content-type. Content type parameter for binary body. @@ -236826,108 +249363,7 @@ async def update_knowledge_base( # response body for status code(s): 200 response == { - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "eval_run_uuid": "str" # Optional. UUID of the created evaluation run. } # response body for status code(s): 404 response == { @@ -236943,16 +249379,14 @@ async def update_knowledge_base( """ @distributed_trace_async - async def update_knowledge_base( - self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + async def create_model_evaluation_run( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Update a Knowledge Base. + """Create Model Evaluation Run. - To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To create a model evaluation run, send a POST request to ``/v2/genai/model_evaluation_runs``. - :param uuid: Knowledge base id. Required. - :type uuid: str :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] :return: JSON object @@ -236964,128 +249398,55 @@ async def update_knowledge_base( # JSON input template you can fill out and use as your body input. body = { - "database_id": "str", # Optional. The id of the DigitalOcean database this - knowledge base will use, optiona. - "embedding_model_uuid": "str", # Optional. Identifier for the foundation - model. - "name": "str", # Optional. Knowledge base name. - "project_id": "str", # Optional. The id of the DigitalOcean project this - knowledge base will belong to. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled for - retrieval. - "model": "str" # Optional. Reranker model internal name. + "candidate_inference_config": { + "max_tokens": 0, # Optional. Inference configuration for the + candidate model during evaluation. + "stop_token": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "system_prompt": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "temperature": 0.0 # Optional. Inference configuration for the + candidate model during evaluation. }, - "tags": [ - "str" # Optional. Tags to organize your knowledge base. + "candidate_model_name": "str", # Optional. Model slug used to call the + candidate model API. For dedicated inference, this is the model slug from the + deployment. For serverless, this should match the model's internal name. + "candidate_model_source": "CANDIDATE_MODEL_SOURCE_SERVERLESS", # Optional. + Default value is "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether inference runs + against the serverless platform, a dedicated deployment, or a model router. Known + values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. UUID of the candidate model to + evaluate. + "dataset_uuid": "str", # Optional. UUID of the dataset to use for + evaluation. + "eval_preset_uuid": "str", # Optional. + "judge_model_uuid": "str", # Optional. UUID of the judge model used to score + responses. + "metric_uuids": [ + "str" # Optional. UUIDs of metrics to evaluate (selected from + ListModelEvaluationMetrics). ], - "uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. + "preset_name": "str", # Optional. + "save_as_preset": bool, # Optional. .. role:: raw-html-m2r(raw) :format: + html If true, saves the inline config as a reusable preset"" + :raw-html-m2r:`
` Ignored when eval_preset_uuid is provided. + "source": "str", # Optional. Source of the run creation (api, sdk, cli). + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + } } # response body for status code(s): 200 response == { - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "eval_run_uuid": "str" # Optional. UUID of the created evaluation run. } # response body for status code(s): 404 response == { @@ -237131,8 +249492,7 @@ async def update_knowledge_base( else: _json = None - _request = build_genai_update_knowledge_base_request( - uuid=uuid, + _request = build_genai_create_model_evaluation_run_request( content_type=content_type, json=_json, content=_content, @@ -237195,14 +249555,27 @@ async def update_knowledge_base( return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + async def get_model_evaluation_run( + self, + eval_run_uuid: str, + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any + ) -> JSON: # pylint: disable=line-too-long - """Delete a Knowledge Base. + """Retrieve Model Evaluation Run. - To delete a knowledge base, send a DELETE request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To retrieve a model evaluation run, send a GET request to + ``/v2/genai/model_evaluation_runs/{eval_run_uuid}``. - :param uuid: Knowledge base id. Required. - :type uuid: str + :param eval_run_uuid: UUID of the evaluation run. Required. + :type eval_run_uuid: str + :keyword page: Page number for per-prompt results (defaults to 1). Default value is None. + :paramtype page: int + :keyword per_page: Number of per-prompt results per page (defaults to 50). Default value is + None. + :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -237212,7 +249585,323 @@ async def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "uuid": "str" # Optional. The id of the deleted knowledge base. + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "results": [ + { + "candidate_model_name": "str", # Optional. Paginated + per-prompt evaluation results. + "candidate_model_uuid": "str", # Optional. Paginated + per-prompt evaluation results. + "ground_truth": "str", # Optional. Paginated per-prompt + evaluation results. + "input": "str", # Optional. The input query sent to the + candidate model. + "metric_results": [ + { + "error_description": "str", # Optional. + Error description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric + name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value + of the metric as a number. + "reasoning": "str", # Optional. Reasoning of + the metric result. + "string_value": "str" # Optional. The value + of the metric as a string. + } + ], + "output": "str" # Optional. The response from the candidate + model. + } + ], + "run": { + "candidate_inference_config": { + "max_tokens": 0, # Optional. Inference configuration for the + candidate model during evaluation. + "stop_token": "str", # Optional. Inference configuration for + the candidate model during evaluation. + "system_prompt": "str", # Optional. Inference configuration + for the candidate model during evaluation. + "temperature": 0.0 # Optional. Inference configuration for + the candidate model during evaluation. + }, + "candidate_model_name": "str", # Optional. Model Evaluation Run + Detail - full view returned when fetching a specific run. + "candidate_model_source": "CANDIDATE_MODEL_SOURCE_SERVERLESS", # + Optional. Default value is "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether + inference runs against the serverless platform, a dedicated deployment, or a + model router. Known values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. Candidate model being + evaluated. + "completed_at": "2020-02-20 00:00:00", # Optional. Model Evaluation + Run Detail - full view returned when fetching a specific run. + "created_at": "2020-02-20 00:00:00", # Optional. Model Evaluation + Run Detail - full view returned when fetching a specific run. + "dataset_name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "dataset_uuid": "str", # Optional. Dataset used for the evaluation. + "error_description": "str", # Optional. Error description if the run + failed or partially succeeded. + "eval_preset_name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "eval_preset_uuid": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "eval_run_uuid": "str", # Optional. UUID of the evaluation run. + "judge_model_name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "judge_model_uuid": "str", # Optional. Judge model used to score + responses. + "metrics": [ + { + "category": "METRIC_CATEGORY_UNSPECIFIED", # + Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known + values are: "METRIC_CATEGORY_UNSPECIFIED", + "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", + "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. Metrics selected + for this evaluation. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", + # Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope + that determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to + agent metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. Metrics selected + for this evaluation. + "metric_name": "str", # Optional. Metrics selected + for this evaluation. + "metric_rank": 0, # Optional. Metrics selected for + this evaluation. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values + are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. Metrics selected + for this evaluation. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for + the metric. + "range_min": 0.0 # Optional. The minimum value for + the metric. + } + ], + "name": "str", # Optional. Name of the evaluation run. + "result_summary": { + "end_time": "2020-02-20 00:00:00", # Optional. Aggregated + result summary for a completed model evaluation run. + "metric_summaries": [ + { + "description": "str", # Optional. Per-metric + aggregated pass/fail statistics. + "fail_percent": 0.0, # Optional. Per-metric + aggregated pass/fail statistics. + "metric_name": "str", # Optional. Per-metric + aggregated pass/fail statistics. + "metric_uuid": "str", # Optional. Per-metric + aggregated pass/fail statistics. + "pass_percent": 0.0 # Optional. Per-metric + aggregated pass/fail statistics. + } + ], + "overall_score_percent": 0.0, # Optional. Aggregated result + summary for a completed model evaluation run. + "per_model_summaries": { + "summaries": [ + { + "metric_summaries": [ + { + "description": "str", + # Optional. Per-metric pass/fail for only this + model's prompts. + "fail_percent": 0.0, + # Optional. Per-metric pass/fail for only this + model's prompts. + "metric_name": "str", + # Optional. Per-metric pass/fail for only this + model's prompts. + "metric_uuid": "str", + # Optional. Per-metric pass/fail for only this + model's prompts. + "pass_percent": 0.0 + # Optional. Per-metric pass/fail for only this + model's prompts. + } + ], + "model_name": "str", # Optional. + Name/slug of the model (matches routed_model from results). + "performance_metrics": { + "candidate_latency": { + "avg_e2e_latency_ms": + 0.0, # Optional. Average end-to-end latency across + all invocations. + "max_e2e_latency_ms": + 0.0, # Optional. Maximum end-to-end latency + observed. + "min_e2e_latency_ms": + 0.0, # Optional. Minimum end-to-end latency + observed. + "p50_latency_ms": + 0.0, # Optional. P50 (median) latency. + "p90_latency_ms": + 0.0, # Optional. P90 latency. + "p95_latency_ms": 0.0 + # Optional. P95 latency. + }, + "token_usage": { + "total_candidate_input_tokens": "str", # Optional. + All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_output_tokens": "str", # Optional. + All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_tokens": "str", # Optional. All + performance metrics are for the candidate model + unless noted otherwise. + "total_judge_input_tokens": "str", # Optional. All + performance metrics are for the candidate model + unless noted otherwise. + "total_judge_output_tokens": "str", # Optional. All + performance metrics are for the candidate model + unless noted otherwise. + "total_judge_tokens": + "str" # Optional. All performance metrics are for + the candidate model unless noted otherwise. + } + }, + "prompt_count": 0 # Optional. Number + of prompts routed to this model. + } + ] + }, + "performance_metrics": { + "candidate_latency": { + "avg_e2e_latency_ms": 0.0, # Optional. + Average end-to-end latency across all invocations. + "max_e2e_latency_ms": 0.0, # Optional. + Maximum end-to-end latency observed. + "min_e2e_latency_ms": 0.0, # Optional. + Minimum end-to-end latency observed. + "p50_latency_ms": 0.0, # Optional. P50 + (median) latency. + "p90_latency_ms": 0.0, # Optional. P90 + latency. + "p95_latency_ms": 0.0 # Optional. P95 + latency. + }, + "token_usage": { + "total_candidate_input_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_output_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_tokens": "str", # Optional. + All performance metrics are for the candidate model unless noted + otherwise. + "total_judge_input_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_judge_output_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_judge_tokens": "str" # Optional. All + performance metrics are for the candidate model unless noted + otherwise. + } + }, + "pricing": { + "currency": "str", # Optional. Currency code (e.g., + "USD"). + "judge_model_pricing": { + "input_cost": 0.0, # Optional. Cost of input + tokens. + "output_cost": 0.0, # Optional. Cost of + output tokens. + "total_cost": 0.0 # Optional. Total cost + (input + output). + }, + "per_candidate_model_pricing": [ + { + "model_name": "str", # Optional. + Model name (for display purposes). + "model_uuid": "str", # Optional. + Model UUID. + "pricing": { + "input_cost": 0.0, # + Optional. Cost of input tokens. + "output_cost": 0.0, # + Optional. Cost of output tokens. + "total_cost": 0.0 # + Optional. Total cost (input + output). + }, + "prompt_count": 0 # Optional. Number + of prompts/rows routed to this model. + } + ], + "total_cost": 0.0 # Optional. Total cost of the + evaluation run (all candidates + judge). + }, + "star_metric_summary": { + "metric_name": "str", # Optional. Star metric + summary with identifying details and threshold. + "metric_uuid": "str", # Optional. Star metric + summary with identifying details and threshold. + "threshold": 0.0 # Optional. Star metric summary + with identifying details and threshold. + }, + "start_time": "2020-02-20 00:00:00", # Optional. Run timing. + "total_duration_seconds": 0 # Optional. Total wall-clock + duration in seconds. + }, + "star_metric": { + "metric_uuid": "str", # Optional. Model Evaluation Run + Detail - full view returned when fetching a specific run. + "name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "success_threshold": 0.0, # Optional. The success threshold + for the star metric. This is a value that the metric must reach to be + considered successful. + "success_threshold_pct": 0 # Optional. The success threshold + for the star metric. This is a percentage value between 0 and 100. + }, + "started_at": "2020-02-20 00:00:00", # Optional. Model Evaluation + Run Detail - full view returned when fetching a specific run. + "status": "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED" # Optional. + Default value is "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED". Model Evaluation + Run Statuses. Known values are: "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", + "MODEL_EVALUATION_RUN_QUEUED", "MODEL_EVALUATION_RUN_RUNNING_DATASET", + "MODEL_EVALUATION_RUN_EVALUATING_RESULTS", "MODEL_EVALUATION_RUN_CANCELLING", + "MODEL_EVALUATION_RUN_CANCELLED", "MODEL_EVALUATION_RUN_SUCCESSFUL", + "MODEL_EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "MODEL_EVALUATION_RUN_FAILED". + } } # response body for status code(s): 404 response == { @@ -237244,8 +249933,125 @@ async def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_delete_knowledge_base_request( - uuid=uuid, + _request = build_genai_get_model_evaluation_run_request( + eval_run_uuid=eval_run_uuid, + page=page, + per_page=per_page, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def get_model_evaluation_run_results_download_url( # pylint: disable=name-too-long + self, eval_run_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Get Download URL for Model Evaluation Run Results. + + To get a presigned download URL for model evaluation run results (gzip-compressed JSON), send a + GET request to ``/v2/genai/model_evaluation_runs/{eval_run_uuid}/results/download_url``. + + :param eval_run_uuid: UUID of the evaluation run. Required. + :type eval_run_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "download_url": "str", # Optional. The presigned URL to download the + gzip-compressed JSON results file (.json.gz). + "expires_at": "2020-02-20 00:00:00" # Optional. The time the URL expires at. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_model_evaluation_run_results_download_url_request( + eval_run_uuid=eval_run_uuid, headers=_headers, params=_params, ) @@ -237367,8 +250173,28 @@ async def list_models( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. Model capabilities (inference, + reasoning, vectorization, etc.). + ], + "context_window": "str", # Optional. Context window (maximum + tokens). "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "description": "str", # Optional. Model description. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "id": "str", # Optional. Human-readable model identifier. "is_foundational": bool, # Optional. True if it is a foundational model provided by do. @@ -237378,9 +250204,79 @@ async def list_models( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, + "model_availability": "str", # Optional. Model availability + (serverless, dedicated, etc.). "name": "str", # Optional. Display name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. + "pricing": { + "input_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "output_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "price_per_audio": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_image": 0.0, # Optional. Unit-based + pricing for non-token models (e.g., Fal AI image/video/audio + generation, speech models). At most one of these is typically + populated per model. Token-based models (chat, embeddings) leave all + of these at 0 and populate input_price_per_million / + output_price_per_million instead. + "price_per_megapixel": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_second": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_thousand_characters": 0.0, # Optional. + Pricing per million tokens (aligns with existing ModelPrice pattern). + "price_per_video": 0.0 # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + }, + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. + Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: + "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and + "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -238250,15 +251146,305 @@ async def delete_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @distributed_trace_async - async def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: + async def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Regenerate API Key for a Model. + + To regenerate a model API key, send a PUT request to + ``/v2/gen-ai/models/api_keys/{api_key_uuid}/regenerate``. + + :param api_key_uuid: API key ID. Required. + :type api_key_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "api_key_info": { + "created_at": "2020-02-20 00:00:00", # Optional. Creation date. + "created_by": "str", # Optional. Created by. + "deleted_at": "2020-02-20 00:00:00", # Optional. Deleted date. + "name": "str", # Optional. Name. + "secret_key": "str", # Optional. Model API Key Info. + "uuid": "str" # Optional. Uuid. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_regenerate_model_api_key_request( + api_key_uuid=api_key_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def list_model_catalog( + self, *, page: Optional[int] = None, limit: Optional[int] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """List Model Catalog. + + Returns all available models. + + :keyword page: Default value is None. + :paramtype page: int + :keyword limit: Default value is None. + :paramtype limit: int + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "data": [ + { + "availability": [ + "str" # Optional. + ], + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. + ], + "context_window": "str", # Optional. Specs (flat). + "creator": "str", # Optional. Model creator/developer (e.g., + "Meta", "Anthropic", "OpenAI"). + "id": "str", # Optional. Identity. + "model_id": "str", # Optional. Model identifier used for API + calls (e.g., "llama3.1-70b-instruct"). + "name": "str", # Optional. + "parameter_count": 0.0, # Optional. + "pricing": { + "input_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "output_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "price_per_audio": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_image": 0.0, # Optional. Unit-based + pricing for non-token models (e.g., Fal AI image/video/audio + generation, speech models). At most one of these is typically + populated per model. Token-based models (chat, embeddings) leave all + of these at 0 and populate input_price_per_million / + output_price_per_million instead. + "price_per_megapixel": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_second": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_thousand_characters": 0.0, # Optional. + Pricing per million tokens (aligns with existing ModelPrice pattern). + "price_per_video": 0.0 # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + }, + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. + Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: + "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and + "MODEL_PROVIDER_OPENAI". + "short_description": "str", # Optional. + "type": "str" # Optional. + } + ], + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_list_model_catalog_request( + page=page, + limit=limit, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + await response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace_async + async def get_model_catalog_card( + self, id: str, *, model_id: Optional[str] = None, **kwargs: Any + ) -> JSON: # pylint: disable=line-too-long - """Regenerate API Key for a Model. + """Get Model Catalog Card. - To regenerate a model API key, send a PUT request to - ``/v2/gen-ai/models/api_keys/{api_key_uuid}/regenerate``. + Returns detailed information for a specific model in the catalog including capabilities, + pricing, and code examples. - :param api_key_uuid: API key ID. Required. - :type api_key_uuid: str + :param id: Required. + :type id: str + :keyword model_id: Model identifier used for API calls (e.g., "llama3.1-70b-instruct"). + Alternative to UUID lookup. Default value is None. + :paramtype model_id: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -238268,13 +251454,70 @@ async def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JS # response body for status code(s): 200 response == { - "api_key_info": { - "created_at": "2020-02-20 00:00:00", # Optional. Creation date. - "created_by": "str", # Optional. Created by. - "deleted_at": "2020-02-20 00:00:00", # Optional. Deleted date. - "name": "str", # Optional. Name. - "secret_key": "str", # Optional. Model API Key Info. - "uuid": "str" # Optional. Uuid. + "data": { + "availability": [ + "str" # Optional. Detail view for GetModelCatalogCard. + ], + "benchmark_score": {}, # Optional. Benchmark scores for this model, + stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. Detail view for GetModelCatalogCard. + ], + "code_snippets": { + "curl": "str", # Optional. Code examples for using the + model. + "javascript": "str", # Optional. Code examples for using the + model. + "python": "str", # Optional. Code examples for using the + model. + "sdk": "str" # Optional. Code examples for using the model. + }, + "context_window": "str", # Optional. Specs (same as Entry). + "creator": "str", # Optional. Model creator/developer (e.g., "Meta", + "Anthropic", "OpenAI"). + "description": "str", # Optional. Card-specific. + "id": "str", # Optional. Identity (same as Entry). + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, + "model_id": "str", # Optional. Model identifier used for API calls + (e.g., "llama3.1-70b-instruct"). + "name": "str", # Optional. Detail view for GetModelCatalogCard. + "parameter_count": 0.0, # Optional. Detail view for + GetModelCatalogCard. + "pricing": { + "input_price_per_million": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "output_price_per_million": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_audio": 0.0, # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + "price_per_image": 0.0, # Optional. Unit-based pricing for + non-token models (e.g., Fal AI image/video/audio generation, speech + models). At most one of these is typically populated per model. + Token-based models (chat, embeddings) leave all of these at 0 and + populate input_price_per_million / output_price_per_million instead. + "price_per_megapixel": 0.0, # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + "price_per_second": 0.0, # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + "price_per_thousand_characters": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "price_per_video": 0.0 # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + }, + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value + is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: + "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and + "MODEL_PROVIDER_OPENAI". + "short_description": "str", # Optional. Detail view for + GetModelCatalogCard. + "type": "str" # Optional. Detail view for GetModelCatalogCard. } } # response body for status code(s): 404 @@ -238307,8 +251550,9 @@ async def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JS cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_regenerate_model_api_key_request( - api_key_uuid=api_key_uuid, + _request = build_genai_get_model_catalog_card_request( + id=id, + model_id=model_id, headers=_headers, params=_params, ) @@ -238427,8 +251671,7 @@ async def list_model_routers( ], "selection_policy": { "prefer": "str" # Optional. - One of: none, cheapest, fastest. Known values are: - "none", "cheapest", and "fastest". + One of: none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. @@ -238583,7 +251826,7 @@ async def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -238613,8 +251856,7 @@ async def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -238689,8 +251931,7 @@ async def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -238757,7 +251998,7 @@ async def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -238787,8 +252028,7 @@ async def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -238971,8 +252211,7 @@ async def list_model_router_presets( ], "selection_policy": { "prefer": "str" # Optional. - One of: none, cheapest, fastest. Known values are: - "none", "cheapest", and "fastest". + One of: none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. @@ -239132,7 +252371,7 @@ async def list_model_router_task_presets( "name": "str", # Optional. Display name. "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "tags": [ "str" # Optional. Lightweight labels for filtering. @@ -239269,8 +252508,7 @@ async def get_model_router(self, uuid: str, **kwargs: Any) -> JSON: ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -239409,7 +252647,7 @@ async def update_model_router( body = { "description": "str", # Optional. Model router description. "fallback_models": [ - "str" # Optional. Fallback models. + {} # Optional. ], "name": "str", # Optional. Model router name. "policies": [ @@ -239424,7 +252662,7 @@ async def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -239455,8 +252693,7 @@ async def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -239534,8 +252771,7 @@ async def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -239589,7 +252825,7 @@ async def update_model_router( body = { "description": "str", # Optional. Model router description. "fallback_models": [ - "str" # Optional. Fallback models. + {} # Optional. ], "name": "str", # Optional. Model router name. "policies": [ @@ -239604,7 +252840,7 @@ async def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -239635,8 +252871,7 @@ async def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -240268,8 +253503,27 @@ async def list_openai_api_keys( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -240283,16 +253537,64 @@ async def list_openai_api_keys( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -240473,211 +253775,358 @@ async def create_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, - "created_at": "2020-02-20 00:00:00", # Optional. - Creation date / time. - "inference_name": "str", # Optional. Internally used - name. - "inference_version": "str", # Optional. Internally - used version. - "is_foundational": bool, # Optional. True if it is a - foundational model provided by do. - "kb_default_chunk_size": 0, # Optional. Default - chunking size limit to show in UI. - "kb_max_chunk_size": 0, # Optional. Maximum chunk - size limit of model. - "kb_min_chunk_size": 0, # Optional. Minimum chunking - size token limits if model supports KNOWLEDGEBASE usecase. - "metadata": {}, # Optional. Additional meta data. - "name": "str", # Optional. Name of the model. - "parent_uuid": "str", # Optional. Unique id of the - model, this model is based on. - "provider": "MODEL_PROVIDER_DIGITALOCEAN", # - Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known - values are: "MODEL_PROVIDER_DIGITALOCEAN", - "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". - "updated_at": "2020-02-20 00:00:00", # Optional. - Last modified. - "upload_complete": bool, # Optional. Model has been - fully uploaded. - "url": "str", # Optional. Download url. - "usecases": [ - "str" # Optional. Usecases of the model. + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], + "inference_name": "str", # Optional. Internally used + name. + "inference_version": "str", # Optional. Internally + used version. + "is_foundational": bool, # Optional. True if it is a + foundational model provided by do. + "kb_default_chunk_size": 0, # Optional. Default + chunking size limit to show in UI. + "kb_max_chunk_size": 0, # Optional. Maximum chunk + size limit of model. + "kb_min_chunk_size": 0, # Optional. Minimum chunking + size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). + "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, + "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. + "parent_uuid": "str", # Optional. Unique id of the + model, this model is based on. + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # + Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known + values are: "MODEL_PROVIDER_DIGITALOCEAN", + "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "upload_complete": bool, # Optional. Model has been + fully uploaded. + "url": "str", # Optional. Download url. + "usecases": [ + "str" # Optional. Usecases of the model. + ], + "uuid": "str", # Optional. Unique id. + "version": { + "major": 0, # Optional. Major version + number. + "minor": 0, # Optional. Minor version + number. + "patch": 0 # Optional. Patch version number. + } + } + ], + "name": "str", # Optional. Name. + "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated + date. + "uuid": "str" # Optional. Uuid. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + async def create_openai_api_key( + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create OpenAI API Key. + + To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "api_key_info": { + "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. + "created_by": "str", # Optional. Created by user id from DO. + "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. + "models": [ + { + "agreement": { + "description": "str", # Optional. Agreement + Description. + "name": "str", # Optional. Agreement + Description. + "url": "str", # Optional. Agreement + Description. + "uuid": "str" # Optional. Agreement + Description. + }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], + "inference_name": "str", # Optional. Internally used + name. + "inference_version": "str", # Optional. Internally + used version. + "is_foundational": bool, # Optional. True if it is a + foundational model provided by do. + "kb_default_chunk_size": 0, # Optional. Default + chunking size limit to show in UI. + "kb_max_chunk_size": 0, # Optional. Maximum chunk + size limit of model. + "kb_min_chunk_size": 0, # Optional. Minimum chunking + size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). + "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, + "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. + "parent_uuid": "str", # Optional. Unique id of the + model, this model is based on. + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # + Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known + values are: "MODEL_PROVIDER_DIGITALOCEAN", + "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "upload_complete": bool, # Optional. Model has been + fully uploaded. + "url": "str", # Optional. Download url. + "usecases": [ + "str" # Optional. Usecases of the model. + ], + "uuid": "str", # Optional. Unique id. + "version": { + "major": 0, # Optional. Major version + number. + "minor": 0, # Optional. Minor version + number. + "patch": 0 # Optional. Patch version number. + } + } + ], + "name": "str", # Optional. Name. + "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated + date. + "uuid": "str" # Optional. Uuid. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace_async + async def create_openai_api_key( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create OpenAI API Key. + + To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. + + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "api_key": "str", # Optional. OpenAI API key. + "name": "str" # Optional. Name of the key. + } + + # response body for status code(s): 200 + response == { + "api_key_info": { + "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. + "created_by": "str", # Optional. Created by user id from DO. + "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. + "models": [ + { + "agreement": { + "description": "str", # Optional. Agreement + Description. + "name": "str", # Optional. Agreement + Description. + "url": "str", # Optional. Agreement + Description. + "uuid": "str" # Optional. Agreement + Description. + }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). ], - "uuid": "str", # Optional. Unique id. - "version": { - "major": 0, # Optional. Major version - number. - "minor": 0, # Optional. Minor version - number. - "patch": 0 # Optional. Patch version number. - } - } - ], - "name": "str", # Optional. Name. - "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated - date. - "uuid": "str" # Optional. Uuid. - } - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @overload - async def create_openai_api_key( - self, - body: Optional[IO[bytes]] = None, - *, - content_type: str = "application/json", - **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Create OpenAI API Key. - - To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. - - :param body: Default value is None. - :type body: IO[bytes] - :keyword content_type: Body Parameter content-type. Content type parameter for binary body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "api_key_info": { - "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. - "created_by": "str", # Optional. Created by user id from DO. - "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. - "models": [ - { - "agreement": { - "description": "str", # Optional. Agreement - Description. - "name": "str", # Optional. Agreement - Description. - "url": "str", # Optional. Agreement - Description. - "uuid": "str" # Optional. Agreement - Description. - }, + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. - "inference_name": "str", # Optional. Internally used - name. - "inference_version": "str", # Optional. Internally - used version. - "is_foundational": bool, # Optional. True if it is a - foundational model provided by do. - "kb_default_chunk_size": 0, # Optional. Default - chunking size limit to show in UI. - "kb_max_chunk_size": 0, # Optional. Maximum chunk - size limit of model. - "kb_min_chunk_size": 0, # Optional. Minimum chunking - size token limits if model supports KNOWLEDGEBASE usecase. - "metadata": {}, # Optional. Additional meta data. - "name": "str", # Optional. Name of the model. - "parent_uuid": "str", # Optional. Unique id of the - model, this model is based on. - "provider": "MODEL_PROVIDER_DIGITALOCEAN", # - Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known - values are: "MODEL_PROVIDER_DIGITALOCEAN", - "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". - "updated_at": "2020-02-20 00:00:00", # Optional. - Last modified. - "upload_complete": bool, # Optional. Model has been - fully uploaded. - "url": "str", # Optional. Download url. - "usecases": [ - "str" # Optional. Usecases of the model. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } ], - "uuid": "str", # Optional. Unique id. - "version": { - "major": 0, # Optional. Major version - number. - "minor": 0, # Optional. Minor version - number. - "patch": 0 # Optional. Patch version number. - } - } - ], - "name": "str", # Optional. Name. - "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated - date. - "uuid": "str" # Optional. Uuid. - } - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @distributed_trace_async - async def create_openai_api_key( - self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Create OpenAI API Key. - - To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. - - :param body: Is either a JSON type or a IO[bytes] type. Default value is None. - :type body: JSON or IO[bytes] - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "api_key": "str", # Optional. OpenAI API key. - "name": "str" # Optional. Name of the key. - } - - # response body for status code(s): 200 - response == { - "api_key_info": { - "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. - "created_by": "str", # Optional. Created by user id from DO. - "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. - "models": [ - { - "agreement": { - "description": "str", # Optional. Agreement - Description. - "name": "str", # Optional. Agreement - Description. - "url": "str", # Optional. Agreement - Description. - "uuid": "str" # Optional. Agreement - Description. - }, - "created_at": "2020-02-20 00:00:00", # Optional. - Creation date / time. "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -240690,14 +254139,59 @@ async def create_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -240863,8 +254357,27 @@ async def get_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -240877,14 +254390,59 @@ async def get_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -241052,8 +254610,27 @@ async def update_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -241066,14 +254643,59 @@ async def update_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -241157,8 +254779,27 @@ async def update_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -241171,14 +254812,59 @@ async def update_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -241264,8 +254950,27 @@ async def update_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -241278,14 +254983,59 @@ async def update_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -241452,8 +255202,27 @@ async def delete_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -241466,14 +255235,59 @@ async def delete_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -241904,6 +255718,24 @@ async def list_agents_by_openai_key( stream. }, "max_tokens": 0, # Optional. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -241915,8 +255747,27 @@ async def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -241929,14 +255780,59 @@ async def list_agents_by_openai_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -241974,8 +255870,27 @@ async def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -241989,10 +255904,25 @@ async def list_agents_by_openai_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -242000,6 +255930,42 @@ async def list_agents_by_openai_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -242030,6 +255996,45 @@ async def list_agents_by_openai_key( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -242050,8 +256055,27 @@ async def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -242065,10 +256089,25 @@ async def list_agents_by_openai_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -242076,6 +256115,42 @@ async def list_agents_by_openai_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -242109,6 +256184,8 @@ async def list_agents_by_openai_key( "project_id": "str", # Optional. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -242289,8 +256366,27 @@ async def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -242304,16 +256400,64 @@ async def list_agents_by_openai_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -242357,6 +256501,8 @@ async def list_agents_by_openai_key( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -242399,6 +256545,13 @@ async def list_agents_by_openai_key( 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -242429,6 +256582,16 @@ async def list_agents_by_openai_key( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -242442,8 +256605,10 @@ async def list_agents_by_openai_key( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -243646,6 +257811,27 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: Name of the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. + Optional list of allowed tool names to expose from + this server. + ], + "authorization": "str", # + Optional. Optional authorization header value for the MCP + server. + "headers": { + "str": "str" # + Optional. Optional additional headers to send to the + MCP server. + }, + "server_label": "str", # + Optional. A label identifying this MCP server. + "server_url": "str" # + Optional. The URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # @@ -243657,8 +257843,27 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -243672,10 +257877,25 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -243683,6 +257903,42 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -243725,9 +257981,30 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": + {}, # Optional. Benchmark scores for this model, + stored as arbitrary JSON. + "capabilities": [ + "str" # + Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": + "str", # Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities supported + by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": @@ -243742,10 +258019,26 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": + "str", # Optional. Lifecycle status of the model + (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" + # Optional. Input/output modalities. + ], + "output": [ + "str" + # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": + 0.0, # Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. @@ -243755,6 +258048,43 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": + [ + "str" # + Optional. Available reasoning efforts for this + model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended + thinking (Anthropic models). + "type": "str", # + Optional. Model type (chat, embedding, image, + reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": @@ -243787,6 +258117,46 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" # Optional. Task name. + }, + "models": [ + "str" + # Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. One of: none, + cheapest, fastest. + }, + "task_slug": + "str" # Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", + # Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of + the model router. + "regions": [ + "str" # Optional. Target + regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", + # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", @@ -243807,9 +258177,30 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": + {}, # Optional. Benchmark scores for this model, + stored as arbitrary JSON. + "capabilities": [ + "str" # + Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": + "str", # Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities supported + by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": @@ -243824,10 +258215,26 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": + "str", # Optional. Lifecycle status of the model + (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" + # Optional. Input/output modalities. + ], + "output": [ + "str" + # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": + 0.0, # Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. @@ -243837,6 +258244,43 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": + [ + "str" # + Optional. Available reasoning efforts for this + model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended + thinking (Anthropic models). + "type": "str", # + Optional. Model type (chat, embedding, image, + reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": @@ -243870,6 +258314,8 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The + reasoning effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is @@ -244070,8 +258516,31 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -244086,10 +258555,25 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -244097,6 +258581,44 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -244143,6 +258665,9 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: # Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The + thinking token budget for Anthropic extended thinking (0 = + disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -244185,6 +258710,13 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -244215,6 +258747,15 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -244228,7 +258769,8 @@ async def list_workspaces(self, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -244704,6 +259246,24 @@ async def create_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -244715,8 +259275,27 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -244730,16 +259309,64 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -244780,8 +259407,31 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -244796,10 +259446,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -244807,6 +259472,44 @@ async def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -244837,6 +259540,48 @@ async def create_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -244857,8 +259602,31 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -244873,10 +259641,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -244884,6 +259667,44 @@ async def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -244917,6 +259738,8 @@ async def create_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -245108,8 +259931,27 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -245123,10 +259965,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -245134,6 +259991,42 @@ async def create_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -245180,6 +260073,8 @@ async def create_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -245219,6 +260114,12 @@ async def create_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -245245,6 +260146,14 @@ async def create_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -245257,8 +260166,9 @@ async def create_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -245644,6 +260554,24 @@ async def create_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -245655,8 +260583,27 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -245670,16 +260617,64 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -245720,8 +260715,31 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -245736,10 +260754,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -245747,6 +260780,44 @@ async def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -245777,6 +260848,48 @@ async def create_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -245797,8 +260910,31 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -245813,10 +260949,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -245824,6 +260975,44 @@ async def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -245857,6 +261046,8 @@ async def create_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -246048,8 +261239,27 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -246063,10 +261273,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -246074,6 +261299,42 @@ async def create_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -246120,6 +261381,8 @@ async def create_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -246159,6 +261422,12 @@ async def create_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -246185,6 +261454,14 @@ async def create_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -246197,8 +261474,9 @@ async def create_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -246586,6 +261864,24 @@ async def create_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -246597,8 +261893,27 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -246612,16 +261927,64 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -246662,8 +262025,31 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -246678,10 +262064,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -246689,6 +262090,44 @@ async def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -246719,6 +262158,48 @@ async def create_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -246739,8 +262220,31 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -246755,10 +262259,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -246766,6 +262285,44 @@ async def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -246799,6 +262356,8 @@ async def create_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -246990,8 +262549,27 @@ async def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -247005,10 +262583,25 @@ async def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -247016,6 +262609,42 @@ async def create_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -247062,6 +262691,8 @@ async def create_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -247101,6 +262732,12 @@ async def create_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -247127,6 +262764,14 @@ async def create_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -247139,8 +262784,9 @@ async def create_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -247610,6 +263256,24 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -247621,8 +263285,27 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -247636,16 +263319,64 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -247686,8 +263417,31 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -247702,10 +263456,25 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -247713,6 +263482,44 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -247743,6 +263550,48 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -247763,8 +263612,31 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -247779,10 +263651,25 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -247790,6 +263677,44 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -247823,6 +263748,8 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -248014,8 +263941,27 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -248029,10 +263975,25 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -248040,6 +264001,42 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -248086,6 +264083,8 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -248125,6 +264124,12 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -248151,6 +264156,14 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -248163,8 +264176,9 @@ async def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -248637,6 +264651,24 @@ async def update_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -248648,8 +264680,27 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -248663,16 +264714,64 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -248713,8 +264812,31 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -248729,10 +264851,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -248740,6 +264877,44 @@ async def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -248770,6 +264945,48 @@ async def update_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -248790,8 +265007,31 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -248806,10 +265046,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -248817,6 +265072,44 @@ async def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -248850,6 +265143,8 @@ async def update_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -249041,8 +265336,27 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -249056,10 +265370,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -249067,6 +265396,42 @@ async def update_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -249113,6 +265478,8 @@ async def update_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -249152,6 +265519,12 @@ async def update_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -249178,6 +265551,14 @@ async def update_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -249190,8 +265571,9 @@ async def update_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -249580,6 +265962,24 @@ async def update_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -249591,8 +265991,27 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -249606,16 +266025,64 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -249656,8 +266123,31 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -249672,10 +266162,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -249683,6 +266188,44 @@ async def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -249713,6 +266256,48 @@ async def update_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -249733,8 +266318,31 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -249749,10 +266357,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -249760,6 +266383,44 @@ async def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -249793,6 +266454,8 @@ async def update_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -249984,8 +266647,27 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -249999,10 +266681,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -250010,6 +266707,42 @@ async def update_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -250056,6 +266789,8 @@ async def update_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -250095,6 +266830,12 @@ async def update_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -250121,6 +266862,14 @@ async def update_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -250133,8 +266882,9 @@ async def update_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -250525,6 +267275,24 @@ async def update_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -250536,8 +267304,27 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -250551,16 +267338,64 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -250601,8 +267436,31 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -250617,10 +267475,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -250628,6 +267501,44 @@ async def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -250658,6 +267569,48 @@ async def update_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -250678,8 +267631,31 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -250694,10 +267670,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -250705,6 +267696,44 @@ async def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -250738,6 +267767,8 @@ async def update_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -250929,8 +267960,27 @@ async def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -250944,10 +267994,25 @@ async def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -250955,6 +268020,42 @@ async def update_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -251001,6 +268102,8 @@ async def update_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -251040,6 +268143,12 @@ async def update_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -251066,6 +268175,14 @@ async def update_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -251078,8 +268195,9 @@ async def update_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -251659,6 +268777,24 @@ async def list_agents_by_workspace( stream. }, "max_tokens": 0, # Optional. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -251670,8 +268806,27 @@ async def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -251684,14 +268839,59 @@ async def list_agents_by_workspace( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -251729,8 +268929,27 @@ async def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -251744,10 +268963,25 @@ async def list_agents_by_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -251755,6 +268989,42 @@ async def list_agents_by_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -251785,6 +269055,45 @@ async def list_agents_by_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -251805,8 +269114,27 @@ async def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -251820,10 +269148,25 @@ async def list_agents_by_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -251831,6 +269174,42 @@ async def list_agents_by_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -251864,6 +269243,8 @@ async def list_agents_by_workspace( "project_id": "str", # Optional. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -252044,8 +269425,27 @@ async def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -252059,16 +269459,64 @@ async def list_agents_by_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -252112,6 +269560,8 @@ async def list_agents_by_workspace( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -252154,6 +269604,13 @@ async def list_agents_by_workspace( 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -252184,6 +269641,16 @@ async def list_agents_by_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -252197,8 +269664,10 @@ async def list_agents_by_workspace( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -252698,6 +270167,24 @@ async def update_agents_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -252709,8 +270196,27 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -252724,16 +270230,64 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -252774,8 +270328,31 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -252790,10 +270367,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -252801,6 +270393,44 @@ async def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -252831,6 +270461,48 @@ async def update_agents_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -252851,8 +270523,31 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -252867,10 +270562,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -252878,6 +270588,44 @@ async def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -252911,6 +270659,8 @@ async def update_agents_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -253102,8 +270852,27 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -253117,10 +270886,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -253128,6 +270912,42 @@ async def update_agents_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -253174,6 +270994,8 @@ async def update_agents_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -253213,6 +271035,12 @@ async def update_agents_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -253239,6 +271067,14 @@ async def update_agents_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -253251,8 +271087,9 @@ async def update_agents_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -253641,6 +271478,24 @@ async def update_agents_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -253652,8 +271507,27 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -253667,16 +271541,64 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -253717,8 +271639,31 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -253733,10 +271678,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -253744,6 +271704,44 @@ async def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -253774,6 +271772,48 @@ async def update_agents_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -253794,8 +271834,31 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -253810,10 +271873,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -253821,6 +271899,44 @@ async def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -253854,6 +271970,8 @@ async def update_agents_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -254045,8 +272163,27 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -254060,10 +272197,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -254071,6 +272223,42 @@ async def update_agents_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -254117,6 +272305,8 @@ async def update_agents_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -254156,6 +272346,12 @@ async def update_agents_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -254182,6 +272378,14 @@ async def update_agents_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -254194,8 +272398,9 @@ async def update_agents_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -254587,6 +272792,24 @@ async def update_agents_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -254598,8 +272821,27 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -254613,16 +272855,64 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -254663,8 +272953,31 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -254679,10 +272992,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -254690,6 +273018,44 @@ async def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -254720,6 +273086,48 @@ async def update_agents_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -254740,8 +273148,31 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -254756,10 +273187,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -254767,6 +273213,44 @@ async def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -254800,6 +273284,8 @@ async def update_agents_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -254991,8 +273477,27 @@ async def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -255006,10 +273511,25 @@ async def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -255017,6 +273537,42 @@ async def update_agents_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -255063,6 +273619,8 @@ async def update_agents_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -255102,6 +273660,12 @@ async def update_agents_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -255128,6 +273692,14 @@ async def update_agents_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -255140,8 +273712,9 @@ async def update_agents_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -255323,6 +273896,11 @@ async def list_evaluation_test_cases_by_workspace( Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # + Optional. Default value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known + values are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of the @@ -255346,6 +273924,14 @@ async def list_evaluation_test_cases_by_workspace( "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether a + metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. "is_metric_goal": bool, # Optional. @@ -255354,7 +273940,8 @@ async def list_evaluation_test_cases_by_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is diff --git a/src/pydo/operations/_operations.py b/src/pydo/operations/_operations.py index e5994cb6..0cf7f791 100644 --- a/src/pydo/operations/_operations.py +++ b/src/pydo/operations/_operations.py @@ -13385,6 +13385,105 @@ def build_genai_list_agents_by_anthropic_key_request( # pylint: disable=name-to ) +def build_genai_list_custom_models_request( + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + status: str = "STATUS_UNSPECIFIED", + **kwargs: Any, +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/custom_models" + + # Construct parameters + if page is not None: + _params["page"] = _SERIALIZER.query("page", page, "int") + if per_page is not None: + _params["per_page"] = _SERIALIZER.query("per_page", per_page, "int") + if status is not None: + _params["status"] = _SERIALIZER.query("status", status, "str") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest( + method="GET", url=_url, params=_params, headers=_headers, **kwargs + ) + + +def build_genai_import_custom_model_request(**kwargs: Any) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/custom_models/import" + + # Construct headers + if content_type is not None: + _headers["Content-Type"] = _SERIALIZER.header( + "content_type", content_type, "str" + ) + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="POST", url=_url, headers=_headers, **kwargs) + + +def build_genai_delete_custom_model_request(uuid: str, **kwargs: Any) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/custom_models/{uuid}" + path_format_arguments = { + "uuid": _SERIALIZER.url("uuid", uuid, "str"), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="DELETE", url=_url, headers=_headers, **kwargs) + + +def build_genai_update_custom_model_metadata_request( # pylint: disable=name-too-long + uuid: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/custom_models/{uuid}/metadata" + path_format_arguments = { + "uuid": _SERIALIZER.url("uuid", uuid, "str"), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct headers + if content_type is not None: + _headers["Content-Type"] = _SERIALIZER.header( + "content_type", content_type, "str" + ) + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="PATCH", url=_url, headers=_headers, **kwargs) + + def build_genai_create_evaluation_dataset_request( **kwargs: Any, ) -> HttpRequest: # pylint: disable=name-too-long @@ -13431,6 +13530,27 @@ def build_genai_create_evaluation_dataset_file_upload_presigned_urls_request( # return HttpRequest(method="POST", url=_url, headers=_headers, **kwargs) +def build_genai_get_evaluation_dataset_download_url_request( # pylint: disable=name-too-long + dataset_uuid: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/evaluation_datasets/{dataset_uuid}/download_url" + path_format_arguments = { + "dataset_uuid": _SERIALIZER.url("dataset_uuid", dataset_uuid, "str"), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="GET", url=_url, headers=_headers, **kwargs) + + def build_genai_list_evaluation_metrics_request( **kwargs: Any, ) -> HttpRequest: # pylint: disable=name-too-long @@ -14114,6 +14234,159 @@ def build_genai_delete_knowledge_base_request( return HttpRequest(method="DELETE", url=_url, headers=_headers, **kwargs) +def build_genai_create_model_eval_dataset_upload_presigned_urls_request( # pylint: disable=name-too-long + **kwargs: Any, +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/model_evaluation/datasets/file_upload_presigned_urls" + + # Construct headers + if content_type is not None: + _headers["Content-Type"] = _SERIALIZER.header( + "content_type", content_type, "str" + ) + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="POST", url=_url, headers=_headers, **kwargs) + + +def build_genai_list_model_evaluation_metrics_request( + **kwargs: Any, +) -> HttpRequest: # pylint: disable=name-too-long + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/model_evaluation_metrics" + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="GET", url=_url, headers=_headers, **kwargs) + + +def build_genai_list_model_evaluation_runs_request( # pylint: disable=name-too-long + *, + eval_preset_uuid: Optional[str] = None, + status: str = "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any, +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/model_evaluation_runs" + + # Construct parameters + if eval_preset_uuid is not None: + _params["eval_preset_uuid"] = _SERIALIZER.query( + "eval_preset_uuid", eval_preset_uuid, "str" + ) + if status is not None: + _params["status"] = _SERIALIZER.query("status", status, "str") + if page is not None: + _params["page"] = _SERIALIZER.query("page", page, "int") + if per_page is not None: + _params["per_page"] = _SERIALIZER.query("per_page", per_page, "int") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest( + method="GET", url=_url, params=_params, headers=_headers, **kwargs + ) + + +def build_genai_create_model_evaluation_run_request( + **kwargs: Any, +) -> HttpRequest: # pylint: disable=name-too-long + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/model_evaluation_runs" + + # Construct headers + if content_type is not None: + _headers["Content-Type"] = _SERIALIZER.header( + "content_type", content_type, "str" + ) + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="POST", url=_url, headers=_headers, **kwargs) + + +def build_genai_get_model_evaluation_run_request( # pylint: disable=name-too-long + eval_run_uuid: str, + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any, +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/model_evaluation_runs/{eval_run_uuid}" + path_format_arguments = { + "eval_run_uuid": _SERIALIZER.url("eval_run_uuid", eval_run_uuid, "str"), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct parameters + if page is not None: + _params["page"] = _SERIALIZER.query("page", page, "int") + if per_page is not None: + _params["per_page"] = _SERIALIZER.query("per_page", per_page, "int") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest( + method="GET", url=_url, params=_params, headers=_headers, **kwargs + ) + + +def build_genai_get_model_evaluation_run_results_download_url_request( # pylint: disable=name-too-long + eval_run_uuid: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/model_evaluation_runs/{eval_run_uuid}/results/download_url" + path_format_arguments = { + "eval_run_uuid": _SERIALIZER.url("eval_run_uuid", eval_run_uuid, "str"), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="GET", url=_url, headers=_headers, **kwargs) + + def build_genai_list_models_request( *, usecases: Optional[List[str]] = None, @@ -14264,6 +14537,59 @@ def build_genai_regenerate_model_api_key_request( # pylint: disable=name-too-lo return HttpRequest(method="PUT", url=_url, headers=_headers, **kwargs) +def build_genai_list_model_catalog_request( + *, page: Optional[int] = None, limit: Optional[int] = None, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/models/catalog" + + # Construct parameters + if page is not None: + _params["page"] = _SERIALIZER.query("page", page, "int") + if limit is not None: + _params["limit"] = _SERIALIZER.query("limit", limit, "int") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest( + method="GET", url=_url, params=_params, headers=_headers, **kwargs + ) + + +def build_genai_get_model_catalog_card_request( # pylint: disable=name-too-long + id: str, *, model_id: Optional[str] = None, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/v2/gen-ai/models/catalog/{id}" + path_format_arguments = { + "id": _SERIALIZER.url("id", id, "str"), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct parameters + if model_id is not None: + _params["model_id"] = _SERIALIZER.query("model_id", model_id, "str") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest( + method="GET", url=_url, params=_params, headers=_headers, **kwargs + ) + + def build_genai_list_model_routers_request( *, page: Optional[int] = None, per_page: Optional[int] = None, **kwargs: Any ) -> HttpRequest: @@ -215048,6 +215374,24 @@ def list_agents( "max_tokens": 0, # Optional. Specifies the maximum number of tokens the model can process in a single input or output, set as a number between 1 and 512. This determines the length of each response. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -215059,8 +215403,27 @@ def list_agents( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -215073,14 +215436,59 @@ def list_agents( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -215098,11 +215506,52 @@ def list_agents( "patch": 0 # Optional. Patch version number. } }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "project_id": "str", # Optional. The DigitalOcean project ID associated with the agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -215287,8 +215736,27 @@ def list_agents( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -215302,16 +215770,64 @@ def list_agents( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -215355,6 +215871,8 @@ def list_agents( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Defines the cumulative probability threshold for word selection, specified as a number between 0 and 1. Higher values allow for more diverse outputs, while lower values ensure @@ -215515,18 +216033,39 @@ def create_agent( "str" # Optional. Ids of the knowledge base(s) to attach to the agent. ], + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI API key ID to use with OpenAI models. "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. + "reasoning_effort": "str", # Optional. "region": "str", # Optional. The DigitalOcean region to deploy your agent in. + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. Agent tag to organize related resources. ], + "thinking_token_budget": 0, # Optional. "workspace_uuid": "str" # Optional. Identifier for the workspace. } @@ -215799,6 +216338,24 @@ def create_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -215807,8 +216364,27 @@ def create_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -215820,14 +216396,56 @@ def create_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -215862,8 +216480,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -215877,16 +216514,64 @@ def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -215915,6 +216600,44 @@ def create_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -215934,8 +216657,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -215949,16 +216691,64 @@ def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -215990,6 +216780,8 @@ def create_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -216167,8 +216959,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -216181,14 +216992,59 @@ def create_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -216228,6 +217084,8 @@ def create_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -216267,6 +217125,13 @@ def create_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -216297,6 +217162,15 @@ def create_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -216310,7 +217184,8 @@ def create_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -216666,6 +217541,24 @@ def create_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -216674,8 +217567,27 @@ def create_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -216687,14 +217599,56 @@ def create_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -216729,8 +217683,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -216744,16 +217717,64 @@ def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -216782,6 +217803,44 @@ def create_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -216801,8 +217860,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -216816,16 +217894,64 @@ def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -216857,6 +217983,8 @@ def create_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -217034,8 +218162,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -217048,14 +218195,59 @@ def create_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -217095,6 +218287,8 @@ def create_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -217134,6 +218328,13 @@ def create_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -217164,6 +218365,15 @@ def create_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -217177,7 +218387,8 @@ def create_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -217271,18 +218482,39 @@ def create_agent( "str" # Optional. Ids of the knowledge base(s) to attach to the agent. ], + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI API key ID to use with OpenAI models. "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. + "reasoning_effort": "str", # Optional. "region": "str", # Optional. The DigitalOcean region to deploy your agent in. + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. Agent tag to organize related resources. ], + "thinking_token_budget": 0, # Optional. "workspace_uuid": "str" # Optional. Identifier for the workspace. } @@ -217555,6 +218787,24 @@ def create_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -217563,8 +218813,27 @@ def create_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -217576,14 +218845,56 @@ def create_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -217618,8 +218929,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -217633,16 +218963,64 @@ def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -217671,6 +219049,44 @@ def create_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -217690,8 +219106,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -217705,16 +219140,64 @@ def create_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -217746,6 +219229,8 @@ def create_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -217923,8 +219408,27 @@ def create_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -217937,14 +219441,59 @@ def create_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -217984,6 +219533,8 @@ def create_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -218023,6 +219574,13 @@ def create_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -218053,6 +219611,15 @@ def create_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -218066,7 +219633,8 @@ def create_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -219455,6 +221023,24 @@ def attach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -219463,8 +221049,27 @@ def attach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -219476,14 +221081,56 @@ def attach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -219518,8 +221165,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -219533,16 +221199,64 @@ def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -219571,6 +221285,44 @@ def attach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -219590,8 +221342,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -219605,16 +221376,64 @@ def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -219646,6 +221465,8 @@ def attach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -219823,8 +221644,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -219837,14 +221677,59 @@ def attach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -219884,6 +221769,8 @@ def attach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -219923,6 +221810,13 @@ def attach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -219953,6 +221847,15 @@ def attach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -219966,7 +221869,8 @@ def attach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -220325,6 +222229,24 @@ def attach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -220333,8 +222255,27 @@ def attach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -220346,14 +222287,56 @@ def attach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -220388,8 +222371,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -220403,16 +222405,64 @@ def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -220441,6 +222491,44 @@ def attach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -220460,8 +222548,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -220475,16 +222582,64 @@ def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -220516,6 +222671,8 @@ def attach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -220693,8 +222850,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -220707,14 +222883,59 @@ def attach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -220754,6 +222975,8 @@ def attach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -220793,6 +223016,13 @@ def attach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -220823,6 +223053,15 @@ def attach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -220836,7 +223075,8 @@ def attach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -221205,6 +223445,24 @@ def attach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -221213,8 +223471,27 @@ def attach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -221226,14 +223503,56 @@ def attach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -221268,8 +223587,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -221283,16 +223621,64 @@ def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -221321,6 +223707,44 @@ def attach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -221340,8 +223764,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -221355,16 +223798,64 @@ def attach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -221396,6 +223887,8 @@ def attach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -221573,8 +224066,27 @@ def attach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -221587,14 +224099,59 @@ def attach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -221634,6 +224191,8 @@ def attach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -221673,6 +224232,13 @@ def attach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -221703,6 +224269,15 @@ def attach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -221716,7 +224291,8 @@ def attach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -222188,6 +224764,24 @@ def update_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -222196,8 +224790,27 @@ def update_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -222209,14 +224822,56 @@ def update_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -222251,8 +224906,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -222266,16 +224940,64 @@ def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -222304,6 +225026,44 @@ def update_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -222323,8 +225083,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -222338,16 +225117,64 @@ def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -222379,6 +225206,8 @@ def update_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -222556,8 +225385,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -222570,14 +225418,59 @@ def update_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -222617,6 +225510,8 @@ def update_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -222656,6 +225551,13 @@ def update_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -222686,6 +225588,15 @@ def update_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -222699,7 +225610,8 @@ def update_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -223061,6 +225973,24 @@ def update_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -223069,8 +225999,27 @@ def update_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -223082,14 +226031,56 @@ def update_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -223124,8 +226115,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -223139,16 +226149,64 @@ def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -223177,6 +226235,44 @@ def update_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -223196,8 +226292,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -223211,16 +226326,64 @@ def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -223252,6 +226415,8 @@ def update_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -223429,8 +226594,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -223443,14 +226627,59 @@ def update_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -223490,6 +226719,8 @@ def update_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -223529,6 +226760,13 @@ def update_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -223559,6 +226797,15 @@ def update_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -223572,7 +226819,8 @@ def update_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -223945,6 +227193,24 @@ def update_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -223953,8 +227219,27 @@ def update_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -223966,14 +227251,56 @@ def update_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -224008,8 +227335,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -224023,16 +227369,64 @@ def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -224061,6 +227455,44 @@ def update_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -224080,8 +227512,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -224095,16 +227546,64 @@ def update_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -224136,6 +227635,8 @@ def update_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -224313,8 +227814,27 @@ def update_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -224327,14 +227847,59 @@ def update_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -224374,6 +227939,8 @@ def update_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -224413,6 +227980,13 @@ def update_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -224443,6 +228017,15 @@ def update_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -224456,7 +228039,8 @@ def update_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -224903,6 +228487,24 @@ def detach_agent_function( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -224911,8 +228513,27 @@ def detach_agent_function( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -224924,14 +228545,56 @@ def detach_agent_function( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -224966,8 +228629,27 @@ def detach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -224981,16 +228663,64 @@ def detach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -225019,6 +228749,44 @@ def detach_agent_function( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -225038,8 +228806,27 @@ def detach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -225053,16 +228840,64 @@ def detach_agent_function( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -225094,6 +228929,8 @@ def detach_agent_function( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -225271,8 +229108,27 @@ def detach_agent_function( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -225285,14 +229141,59 @@ def detach_agent_function( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -225332,6 +229233,8 @@ def detach_agent_function( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -225371,6 +229274,13 @@ def detach_agent_function( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -225401,6 +229311,15 @@ def detach_agent_function( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -225414,7 +229333,8 @@ def detach_agent_function( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -225862,6 +229782,24 @@ def attach_agent_guardrails( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -225870,8 +229808,27 @@ def attach_agent_guardrails( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -225883,14 +229840,56 @@ def attach_agent_guardrails( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -225925,8 +229924,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -225940,16 +229958,64 @@ def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -225978,6 +230044,44 @@ def attach_agent_guardrails( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -225997,8 +230101,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -226012,16 +230135,64 @@ def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -226053,6 +230224,8 @@ def attach_agent_guardrails( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -226230,8 +230403,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -226244,14 +230436,59 @@ def attach_agent_guardrails( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -226291,6 +230528,8 @@ def attach_agent_guardrails( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -226330,6 +230569,13 @@ def attach_agent_guardrails( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -226360,6 +230606,15 @@ def attach_agent_guardrails( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -226373,7 +230628,8 @@ def attach_agent_guardrails( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -226732,6 +230988,24 @@ def attach_agent_guardrails( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -226740,8 +231014,27 @@ def attach_agent_guardrails( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -226753,14 +231046,56 @@ def attach_agent_guardrails( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -226795,8 +231130,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -226810,16 +231164,64 @@ def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -226848,6 +231250,44 @@ def attach_agent_guardrails( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -226867,8 +231307,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -226882,16 +231341,64 @@ def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -226923,6 +231430,8 @@ def attach_agent_guardrails( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -227100,8 +231609,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -227114,14 +231642,59 @@ def attach_agent_guardrails( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -227161,6 +231734,8 @@ def attach_agent_guardrails( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -227200,6 +231775,13 @@ def attach_agent_guardrails( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -227230,6 +231812,15 @@ def attach_agent_guardrails( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -227243,7 +231834,8 @@ def attach_agent_guardrails( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -227608,6 +232200,24 @@ def attach_agent_guardrails( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -227616,8 +232226,27 @@ def attach_agent_guardrails( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -227629,14 +232258,56 @@ def attach_agent_guardrails( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -227671,8 +232342,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -227686,16 +232376,64 @@ def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -227724,6 +232462,44 @@ def attach_agent_guardrails( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -227743,8 +232519,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -227758,16 +232553,64 @@ def attach_agent_guardrails( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -227799,6 +232642,8 @@ def attach_agent_guardrails( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -227976,8 +232821,27 @@ def attach_agent_guardrails( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -227990,14 +232854,59 @@ def attach_agent_guardrails( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -228037,6 +232946,8 @@ def attach_agent_guardrails( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -228076,6 +232987,13 @@ def attach_agent_guardrails( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -228106,6 +233024,15 @@ def attach_agent_guardrails( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -228119,7 +233046,8 @@ def attach_agent_guardrails( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -228564,6 +233492,24 @@ def detach_agent_guardrail( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -228572,8 +233518,27 @@ def detach_agent_guardrail( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -228585,14 +233550,56 @@ def detach_agent_guardrail( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -228627,8 +233634,27 @@ def detach_agent_guardrail( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -228642,16 +233668,64 @@ def detach_agent_guardrail( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -228680,6 +233754,44 @@ def detach_agent_guardrail( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -228699,8 +233811,27 @@ def detach_agent_guardrail( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -228714,16 +233845,64 @@ def detach_agent_guardrail( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -228755,6 +233934,8 @@ def detach_agent_guardrail( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -228932,8 +234113,27 @@ def detach_agent_guardrail( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -228946,14 +234146,59 @@ def detach_agent_guardrail( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -228993,6 +234238,8 @@ def detach_agent_guardrail( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -229032,6 +234279,13 @@ def detach_agent_guardrail( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -229062,6 +234316,15 @@ def detach_agent_guardrail( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -229075,7 +234338,8 @@ def detach_agent_guardrail( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -229500,6 +234764,24 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -229508,8 +234790,27 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -229521,14 +234822,56 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -229563,8 +234906,27 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -229578,16 +234940,64 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -229616,6 +235026,44 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -229635,8 +235083,27 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -229650,16 +235117,64 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -229691,6 +235206,8 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -229868,8 +235385,27 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -229882,14 +235418,59 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -229929,6 +235510,8 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -229968,6 +235551,13 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -229998,6 +235588,15 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -230011,7 +235610,8 @@ def attach_knowledge_bases(self, agent_uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -230439,6 +236039,24 @@ def attach_knowledge_base( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -230447,8 +236065,27 @@ def attach_knowledge_base( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -230460,14 +236097,56 @@ def attach_knowledge_base( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -230502,8 +236181,27 @@ def attach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -230517,16 +236215,64 @@ def attach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -230555,6 +236301,44 @@ def attach_knowledge_base( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -230574,8 +236358,27 @@ def attach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -230589,16 +236392,64 @@ def attach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -230630,6 +236481,8 @@ def attach_knowledge_base( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -230807,8 +236660,27 @@ def attach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -230821,14 +236693,59 @@ def attach_knowledge_base( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -230868,6 +236785,8 @@ def attach_knowledge_base( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -230907,6 +236826,13 @@ def attach_knowledge_base( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -230937,6 +236863,15 @@ def attach_knowledge_base( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -230950,7 +236885,8 @@ def attach_knowledge_base( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -231379,6 +237315,24 @@ def detach_knowledge_base( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -231387,8 +237341,27 @@ def detach_knowledge_base( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -231400,14 +237373,56 @@ def detach_knowledge_base( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -231442,8 +237457,27 @@ def detach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -231457,16 +237491,64 @@ def detach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -231495,6 +237577,44 @@ def detach_knowledge_base( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -231514,8 +237634,27 @@ def detach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -231529,16 +237668,64 @@ def detach_knowledge_base( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -231570,6 +237757,8 @@ def detach_knowledge_base( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -231747,8 +237936,27 @@ def detach_knowledge_base( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -231761,14 +237969,59 @@ def detach_knowledge_base( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -231808,6 +238061,8 @@ def detach_knowledge_base( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -231847,6 +238102,13 @@ def detach_knowledge_base( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -231877,6 +238139,15 @@ def detach_knowledge_base( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -231890,7 +238161,8 @@ def detach_knowledge_base( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -232964,6 +239236,24 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -232972,8 +239262,27 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -232985,14 +239294,56 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -233027,8 +239378,27 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -233042,16 +239412,64 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -233080,6 +239498,44 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -233099,8 +239555,27 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -233114,16 +239589,64 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -233155,6 +239678,8 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -233332,8 +239857,27 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -233346,14 +239890,59 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -233393,6 +239982,8 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -233432,6 +240023,13 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -233462,6 +240060,15 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -233475,7 +240082,8 @@ def get_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -233651,6 +240259,8 @@ def update_agent( ], "anthropic_key_uuid": "str", # Optional. Optional anthropic key uuid for use with anthropic models. + "clear_mcp_servers": bool, # Optional. When true, removes all MCP servers + from the agent. Use this instead of sending an empty mcp_servers array. "conversation_logs_enabled": bool, # Optional. Optional update of conversation logs enabled. "description": "str", # Optional. Agent description. @@ -233663,8 +240273,26 @@ def update_agent( "max_tokens": 0, # Optional. Specifies the maximum number of tokens the model can process in a single input or output, set as a number between 1 and 512. This determines the length of each response. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. Optional Model Provider uuid for use with provider models. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI key uuid for use with @@ -233672,6 +240300,7 @@ def update_agent( "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. "provide_citations": bool, # Optional. + "reasoning_effort": "str", # Optional. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The retrieval method is unknown * RETRIEVAL_METHOD_REWRITE: The retrieval method is rewrite * @@ -233681,12 +240310,14 @@ def update_agent( "RETRIEVAL_METHOD_UNKNOWN", "RETRIEVAL_METHOD_REWRITE", "RETRIEVAL_METHOD_STEP_BACK", "RETRIEVAL_METHOD_SUB_QUERIES", and "RETRIEVAL_METHOD_NONE". + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. A set of abitrary tags to organize your agent. ], "temperature": 0.0, # Optional. Controls the model"u2019s creativity, specified as a number between 0 and 1. Lower values produce more predictable and conservative responses, while higher values encourage creativity and variation. + "thinking_token_budget": 0, # Optional. "top_p": 0.0, # Optional. Defines the cumulative probability threshold for word selection, specified as a number between 0 and 1. Higher values allow for more diverse outputs, while lower values ensure focused and coherent responses. @@ -233962,6 +240593,24 @@ def update_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -233970,8 +240619,27 @@ def update_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -233983,14 +240651,56 @@ def update_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -234025,8 +240735,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -234040,16 +240769,64 @@ def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -234078,6 +240855,44 @@ def update_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -234097,8 +240912,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -234112,16 +240946,64 @@ def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -234153,6 +241035,8 @@ def update_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -234330,8 +241214,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -234344,14 +241247,59 @@ def update_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -234391,6 +241339,8 @@ def update_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -234430,6 +241380,13 @@ def update_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -234460,6 +241417,15 @@ def update_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -234473,7 +241439,8 @@ def update_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -234832,6 +241799,24 @@ def update_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -234840,8 +241825,27 @@ def update_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -234853,14 +241857,56 @@ def update_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -234895,8 +241941,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -234910,16 +241975,64 @@ def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -234948,6 +242061,44 @@ def update_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -234967,8 +242118,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -234982,16 +242152,64 @@ def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -235023,6 +242241,8 @@ def update_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -235200,8 +242420,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -235214,14 +242453,59 @@ def update_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -235261,6 +242545,8 @@ def update_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -235300,6 +242586,13 @@ def update_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -235330,6 +242623,15 @@ def update_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -235343,7 +242645,8 @@ def update_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -235434,6 +242737,8 @@ def update_agent( ], "anthropic_key_uuid": "str", # Optional. Optional anthropic key uuid for use with anthropic models. + "clear_mcp_servers": bool, # Optional. When true, removes all MCP servers + from the agent. Use this instead of sending an empty mcp_servers array. "conversation_logs_enabled": bool, # Optional. Optional update of conversation logs enabled. "description": "str", # Optional. Agent description. @@ -235446,8 +242751,26 @@ def update_agent( "max_tokens": 0, # Optional. Specifies the maximum number of tokens the model can process in a single input or output, set as a number between 1 and 512. This determines the length of each response. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed tool + names to expose from this server. + ], + "authorization": "str", # Optional. Optional authorization + header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional headers + to send to the MCP server. + }, + "server_label": "str", # Optional. A label identifying this + MCP server. + "server_url": "str" # Optional. The URL of the MCP server. + } + ], "model_provider_key_uuid": "str", # Optional. Optional Model Provider uuid for use with provider models. + "model_router_uuid": "str", # Optional. "model_uuid": "str", # Optional. Identifier for the foundation model. "name": "str", # Optional. Agent name. "open_ai_key_uuid": "str", # Optional. Optional OpenAI key uuid for use with @@ -235455,6 +242778,7 @@ def update_agent( "project_id": "str", # Optional. The id of the DigitalOcean project this agent will belong to. "provide_citations": bool, # Optional. + "reasoning_effort": "str", # Optional. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The retrieval method is unknown * RETRIEVAL_METHOD_REWRITE: The retrieval method is rewrite * @@ -235464,12 +242788,14 @@ def update_agent( "RETRIEVAL_METHOD_UNKNOWN", "RETRIEVAL_METHOD_REWRITE", "RETRIEVAL_METHOD_STEP_BACK", "RETRIEVAL_METHOD_SUB_QUERIES", and "RETRIEVAL_METHOD_NONE". + "router_preset_slug": "str", # Optional. "tags": [ "str" # Optional. A set of abitrary tags to organize your agent. ], "temperature": 0.0, # Optional. Controls the model"u2019s creativity, specified as a number between 0 and 1. Lower values produce more predictable and conservative responses, while higher values encourage creativity and variation. + "thinking_token_budget": 0, # Optional. "top_p": 0.0, # Optional. Defines the cumulative probability threshold for word selection, specified as a number between 0 and 1. Higher values allow for more diverse outputs, while lower values ensure focused and coherent responses. @@ -235745,6 +243071,24 @@ def update_agent( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -235753,8 +243097,27 @@ def update_agent( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -235766,14 +243129,56 @@ def update_agent( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -235808,8 +243213,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -235823,16 +243247,64 @@ def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -235861,6 +243333,44 @@ def update_agent( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -235880,8 +243390,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -235895,16 +243424,64 @@ def update_agent( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -235936,6 +243513,8 @@ def update_agent( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -236113,8 +243692,27 @@ def update_agent( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -236127,14 +243725,59 @@ def update_agent( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -236174,6 +243817,8 @@ def update_agent( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -236213,6 +243858,13 @@ def update_agent( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -236243,6 +243895,15 @@ def update_agent( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -236256,7 +243917,8 @@ def update_agent( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -236696,6 +244358,24 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -236704,8 +244384,27 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -236717,14 +244416,56 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -236759,8 +244500,27 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -236774,16 +244534,64 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -236812,6 +244620,44 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -236831,8 +244677,27 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -236846,16 +244711,64 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -236887,6 +244800,8 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -237064,8 +244979,27 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -237078,14 +245012,59 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -237125,6 +245104,8 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -237164,6 +245145,13 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -237194,6 +245182,15 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -237207,7 +245204,8 @@ def delete_agent(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -237653,6 +245651,24 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: stream. }, "max_tokens": 0, # Optional. Child agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -237664,8 +245680,27 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -237678,14 +245713,59 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -237723,8 +245803,27 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -237738,10 +245837,25 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -237749,6 +245863,42 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -237779,6 +245929,45 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -237799,8 +245988,27 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -237814,10 +246022,25 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -237825,6 +246048,42 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -237858,6 +246117,8 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. Child agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -238038,8 +246299,27 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -238053,16 +246333,64 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -238106,6 +246434,8 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Child agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -238148,6 +246478,13 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -238178,6 +246515,16 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -238191,8 +246538,10 @@ def get_agent_children(self, uuid: str, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -238647,6 +246996,24 @@ def update_agent_deployment_visibility( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -238655,8 +247022,27 @@ def update_agent_deployment_visibility( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -238668,14 +247054,56 @@ def update_agent_deployment_visibility( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -238710,8 +247138,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -238725,16 +247172,64 @@ def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -238763,6 +247258,44 @@ def update_agent_deployment_visibility( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -238782,8 +247315,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -238797,16 +247349,64 @@ def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -238838,6 +247438,8 @@ def update_agent_deployment_visibility( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -239015,8 +247617,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -239029,14 +247650,59 @@ def update_agent_deployment_visibility( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -239076,6 +247742,8 @@ def update_agent_deployment_visibility( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -239115,6 +247783,13 @@ def update_agent_deployment_visibility( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -239145,6 +247820,15 @@ def update_agent_deployment_visibility( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -239158,7 +247842,8 @@ def update_agent_deployment_visibility( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -239517,6 +248202,24 @@ def update_agent_deployment_visibility( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -239525,8 +248228,27 @@ def update_agent_deployment_visibility( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -239538,14 +248260,56 @@ def update_agent_deployment_visibility( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -239580,8 +248344,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -239595,16 +248378,64 @@ def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -239633,6 +248464,44 @@ def update_agent_deployment_visibility( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -239652,8 +248521,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -239667,16 +248555,64 @@ def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -239708,6 +248644,8 @@ def update_agent_deployment_visibility( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -239885,8 +248823,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -239899,14 +248856,59 @@ def update_agent_deployment_visibility( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -239946,6 +248948,8 @@ def update_agent_deployment_visibility( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -239985,6 +248989,13 @@ def update_agent_deployment_visibility( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -240015,6 +249026,15 @@ def update_agent_deployment_visibility( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -240028,7 +249048,8 @@ def update_agent_deployment_visibility( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -240393,6 +249414,24 @@ def update_agent_deployment_visibility( "log_stream_name": "str" # Optional. Name of the log stream. }, "max_tokens": 0, # Optional. An Agent. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of allowed + tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional additional + headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of the MCP + server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -240401,8 +249440,27 @@ def update_agent_deployment_visibility( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window size in + tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally used version. @@ -240414,14 +249472,56 @@ def update_agent_deployment_visibility( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -240456,8 +249556,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -240471,16 +249590,64 @@ def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -240509,6 +249676,44 @@ def update_agent_deployment_visibility( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level fallback + models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # + Optional. Short task description. + "name": "str" # Optional. + Task name. + }, + "models": [ + "str" # Optional. Models + assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. + One of: none, cheapest, fastest. + }, + "task_slug": "str" # Optional. Task + slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -240528,8 +249733,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -240543,16 +249767,64 @@ def update_agent_deployment_visibility( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -240584,6 +249856,8 @@ def update_agent_deployment_visibility( "project_id": "str", # Optional. An Agent. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort for the + agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: The @@ -240761,8 +250035,27 @@ def update_agent_deployment_visibility( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -240775,14 +250068,59 @@ def update_agent_deployment_visibility( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -240822,6 +250160,8 @@ def update_agent_deployment_visibility( template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token budget + for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. An Agent. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "url": "str", # Optional. Access your agent under this url. @@ -240861,6 +250201,13 @@ def update_agent_deployment_visibility( # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -240891,6 +250238,15 @@ def update_agent_deployment_visibility( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -240904,7 +250260,8 @@ def update_agent_deployment_visibility( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -242930,6 +252287,24 @@ def list_agents_by_anthropic_key( stream. }, "max_tokens": 0, # Optional. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -242941,8 +252316,27 @@ def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -242955,14 +252349,59 @@ def list_agents_by_anthropic_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -243000,8 +252439,27 @@ def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -243015,10 +252473,25 @@ def list_agents_by_anthropic_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -243026,6 +252499,42 @@ def list_agents_by_anthropic_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -243056,6 +252565,45 @@ def list_agents_by_anthropic_key( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -243076,8 +252624,27 @@ def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -243091,10 +252658,25 @@ def list_agents_by_anthropic_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -243102,6 +252684,42 @@ def list_agents_by_anthropic_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -243135,6 +252753,8 @@ def list_agents_by_anthropic_key( "project_id": "str", # Optional. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -243315,8 +252935,27 @@ def list_agents_by_anthropic_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -243330,16 +252969,64 @@ def list_agents_by_anthropic_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -243383,6 +253070,8 @@ def list_agents_by_anthropic_key( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -243425,6 +253114,13 @@ def list_agents_by_anthropic_key( 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -243455,6 +253151,16 @@ def list_agents_by_anthropic_key( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -243468,8 +253174,10 @@ def list_agents_by_anthropic_key( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -243627,81 +253335,28 @@ def list_agents_by_anthropic_key( return cast(JSON, deserialized) # type: ignore - @overload - def create_evaluation_dataset( - self, - body: Optional[JSON] = None, - *, - content_type: str = "application/json", - **kwargs: Any, - ) -> JSON: - # pylint: disable=line-too-long - """Create Evaluation Dataset. - - To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. - - :param body: Default value is None. - :type body: JSON - :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value - is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: - "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", and - "EVALUATION_DATASET_TYPE_NON_ADK". - "file_upload_dataset": { - "original_file_name": "str", # Optional. The original file name. - "size_in_bytes": "str", # Optional. The size of the file in bytes. - "stored_object_key": "str" # Optional. The object key the file was - stored as. - }, - "name": "str" # Optional. The name of the agent evaluation dataset. - } - - # response body for status code(s): 200 - response == { - "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @overload - def create_evaluation_dataset( + @distributed_trace + def list_custom_models( self, - body: Optional[IO[bytes]] = None, *, - content_type: str = "application/json", + page: Optional[int] = None, + per_page: Optional[int] = None, + status: str = "STATUS_UNSPECIFIED", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Dataset. + """List Custom Models. - To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. + To list custom models, send a GET request to ``/v2/gen-ai/custom_models``. - :param body: Default value is None. - :type body: IO[bytes] - :keyword content_type: Body Parameter content-type. Content type parameter for binary body. - Default value is "application/json". - :paramtype content_type: str + :keyword page: Page number for pagination. Default value is None. + :paramtype page: int + :keyword per_page: Number of items per page. Default value is None. + :paramtype per_page: int + :keyword status: Filter by model status. Known values are: "STATUS_UNSPECIFIED", + "STATUS_IMPORTING", "STATUS_READY", "STATUS_FAILED", and "STATUS_DELETED". Default value is + "STATUS_UNSPECIFIED". + :paramtype status: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -243711,57 +253366,118 @@ def create_evaluation_dataset( # response body for status code(s): 200 response == { - "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @distributed_trace - def create_evaluation_dataset( - self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Create Evaluation Dataset. - - To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. - - :param body: Is either a JSON type or a IO[bytes] type. Default value is None. - :type body: JSON or IO[bytes] - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value - is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: - "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", and - "EVALUATION_DATASET_TYPE_NON_ADK". - "file_upload_dataset": { - "original_file_name": "str", # Optional. The original file name. - "size_in_bytes": "str", # Optional. The size of the file in bytes. - "stored_object_key": "str" # Optional. The object key the file was - stored as. + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } }, - "name": "str" # Optional. The name of the agent evaluation dataset. - } - - # response body for status code(s): 200 - response == { - "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. + "max_threshold": 0, # Optional. Maximum number of custom models allowed for + this team's tier. + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "models": [ + { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 + timestamp indicating when the dedicated inference deployment was + created. + "endpoints": { + "private_endpoint_fqdn": "str", # + Optional. Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # + Optional. Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier + (UUID) of the dedicated inference deployment. + "name": "str", # Optional. Human-readable + name of the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of + the region where the dedicated inference deployment is running + (e.g. "atl1"). + "state": "str", # Optional. Current + lifecycle state of the dedicated inference deployment (e.g. + "ACTIVE", "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 + timestamp indicating when the dedicated inference deployment was + last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from + the model repository. + "context_length": 0, # Optional. Maximum context length + supported by the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly + cost in dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp + when the model was created. + "description": "str", # Optional. Description of the custom + model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., + text, image). + ], + "license": "str", # Optional. License under which the model + is distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., + text, image). + ], + "parameters": "str", # Optional. Number of parameters in the + model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # + Optional. Default value is "ACCESS_TYPE_UNSPECIFIED". Access level + required for the model repository. Known values are: + "ACCESS_TYPE_UNSPECIFIED", "ACCESS_TYPE_PUBLIC", + "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of + the model version. + "hf_token": "str", # Optional. User-provided + HuggingFace token for gated/private models (not persisted in + source_ref). + "prefix": "str", # Optional. Object prefix path in + the bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. + Default value is "SOURCE_TYPE_UNSPECIFIED". Source from which the model + was imported. Known values are: "SOURCE_TYPE_UNSPECIFIED", + "SOURCE_TYPE_HUGGINGFACE", "SOURCE_TYPE_SPACES_BUCKET", + "SOURCE_TYPE_SDK_UPLOAD", and "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. + Known values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", + "STATUS_READY", "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces + bucket where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model + files in bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp + when the model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom + model. + } + ] } # response body for status code(s): 404 response == { @@ -243788,29 +253504,15 @@ def create_evaluation_dataset( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} - content_type: Optional[str] = kwargs.pop( - "content_type", _headers.pop("Content-Type", None) - ) cls: ClsType[JSON] = kwargs.pop("cls", None) - content_type = content_type or "application/json" - _json = None - _content = None - if isinstance(body, (IOBase, bytes)): - _content = body - else: - if body is not None: - _json = body - else: - _json = None - - _request = build_genai_create_evaluation_dataset_request( - content_type=content_type, - json=_json, - content=_content, + _request = build_genai_list_custom_models_request( + page=page, + per_page=per_page, + status=status, headers=_headers, params=_params, ) @@ -243870,7 +253572,7 @@ def create_evaluation_dataset( return cast(JSON, deserialized) # type: ignore @overload - def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + def import_custom_model( self, body: Optional[JSON] = None, *, @@ -243878,10 +253580,9 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Create Presigned URLs for Evaluation Dataset File Upload. + """Import Custom Model. - To create presigned URLs for evaluation dataset file upload, send a POST request to - ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + To import a custom model, send a POST request to ``/v2/gen-ai/custom_models/import``. :param body: Default value is None. :type body: JSON @@ -243897,29 +253598,151 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam # JSON input template you can fill out and use as your body input. body = { - "files": [ - { - "file_name": "str", # Optional. Local filename. - "file_size": "str" # Optional. The size of the file in - bytes. - } - ] + "accept_terms_and_conditions": bool, # Optional. Whether the caller accepts + the terms and conditions for importing this model. + "description": "str", # Optional. Description of the model. + "name": "str", # Optional. Name for the imported model. + "preferred_gpu_region": "str", # Optional. Preferred GPU region for + deployment. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. Default value + is "ACCESS_TYPE_UNSPECIFIED". Access level required for the model repository. + Known values are: "ACCESS_TYPE_UNSPECIFIED", "ACCESS_TYPE_PUBLIC", + "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace token for + gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value is + "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known values + are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + } } # response body for status code(s): 200 response == { - "request_id": "str", # Optional. The ID generated for the request for - Presigned URLs. - "uploads": [ + "error": "str", # Optional. + "import_job": { + "bytes_done": "str", # Optional. Bytes imported so far. + "bytes_total": "str", # Optional. Total bytes to import. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp when + the import completed. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + job was created. + "error_message": "str", # Optional. Error message if import failed. + "error_step": "str", # Optional. Step at which the error occurred. + "files_done": 0, # Optional. Number of files imported so far. + "files_total": 0, # Optional. Total number of files to import. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + import started. + "status": "str", # Optional. Current status of the import job. + "uuid": "str" # Optional. Unique identifier for the import job. + }, + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + }, + "validation_steps": [ { - "expires_at": "2020-02-20 00:00:00", # Optional. The time - the url expires at. - "object_key": "str", # Optional. The unique object key to - store the file as. - "original_file_name": "str", # Optional. The original file - name. - "presigned_url": "str" # Optional. The actual presigned URL - the client can use to upload the file directly. + "error": "str", # Optional. Error message if validation + failed. + "name": "str", # Optional. Name of the validation step. + "passed": bool # Optional. Whether the validation step + passed. } ] } @@ -243937,7 +253760,7 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam """ @overload - def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + def import_custom_model( self, body: Optional[IO[bytes]] = None, *, @@ -243945,10 +253768,9 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Create Presigned URLs for Evaluation Dataset File Upload. + """Import Custom Model. - To create presigned URLs for evaluation dataset file upload, send a POST request to - ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + To import a custom model, send a POST request to ``/v2/gen-ai/custom_models/import``. :param body: Default value is None. :type body: IO[bytes] @@ -243964,18 +253786,117 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam # response body for status code(s): 200 response == { - "request_id": "str", # Optional. The ID generated for the request for - Presigned URLs. - "uploads": [ + "error": "str", # Optional. + "import_job": { + "bytes_done": "str", # Optional. Bytes imported so far. + "bytes_total": "str", # Optional. Total bytes to import. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp when + the import completed. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + job was created. + "error_message": "str", # Optional. Error message if import failed. + "error_step": "str", # Optional. Step at which the error occurred. + "files_done": 0, # Optional. Number of files imported so far. + "files_total": 0, # Optional. Total number of files to import. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + import started. + "status": "str", # Optional. Current status of the import job. + "uuid": "str" # Optional. Unique identifier for the import job. + }, + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + }, + "validation_steps": [ { - "expires_at": "2020-02-20 00:00:00", # Optional. The time - the url expires at. - "object_key": "str", # Optional. The unique object key to - store the file as. - "original_file_name": "str", # Optional. The original file - name. - "presigned_url": "str" # Optional. The actual presigned URL - the client can use to upload the file directly. + "error": "str", # Optional. Error message if validation + failed. + "name": "str", # Optional. Name of the validation step. + "passed": bool # Optional. Whether the validation step + passed. } ] } @@ -243993,14 +253914,13 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam """ @distributed_trace - def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + def import_custom_model( self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Presigned URLs for Evaluation Dataset File Upload. + """Import Custom Model. - To create presigned URLs for evaluation dataset file upload, send a POST request to - ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + To import a custom model, send a POST request to ``/v2/gen-ai/custom_models/import``. :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] @@ -244013,29 +253933,151 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam # JSON input template you can fill out and use as your body input. body = { - "files": [ - { - "file_name": "str", # Optional. Local filename. - "file_size": "str" # Optional. The size of the file in - bytes. - } - ] + "accept_terms_and_conditions": bool, # Optional. Whether the caller accepts + the terms and conditions for importing this model. + "description": "str", # Optional. Description of the model. + "name": "str", # Optional. Name for the imported model. + "preferred_gpu_region": "str", # Optional. Preferred GPU region for + deployment. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. Default value + is "ACCESS_TYPE_UNSPECIFIED". Access level required for the model repository. + Known values are: "ACCESS_TYPE_UNSPECIFIED", "ACCESS_TYPE_PUBLIC", + "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace token for + gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value is + "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known values + are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + } } # response body for status code(s): 200 response == { - "request_id": "str", # Optional. The ID generated for the request for - Presigned URLs. - "uploads": [ + "error": "str", # Optional. + "import_job": { + "bytes_done": "str", # Optional. Bytes imported so far. + "bytes_total": "str", # Optional. Total bytes to import. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp when + the import completed. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + job was created. + "error_message": "str", # Optional. Error message if import failed. + "error_step": "str", # Optional. Step at which the error occurred. + "files_done": 0, # Optional. Number of files imported so far. + "files_total": 0, # Optional. Total number of files to import. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + import started. + "status": "str", # Optional. Current status of the import job. + "uuid": "str" # Optional. Unique identifier for the import job. + }, + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + }, + "validation_steps": [ { - "expires_at": "2020-02-20 00:00:00", # Optional. The time - the url expires at. - "object_key": "str", # Optional. The unique object key to - store the file as. - "original_file_name": "str", # Optional. The original file - name. - "presigned_url": "str" # Optional. The actual presigned URL - the client can use to upload the file directly. + "error": "str", # Optional. Error message if validation + failed. + "name": "str", # Optional. Name of the validation step. + "passed": bool # Optional. Whether the validation step + passed. } ] } @@ -244083,14 +254125,12 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam else: _json = None - _request = ( - build_genai_create_evaluation_dataset_file_upload_presigned_urls_request( - content_type=content_type, - json=_json, - content=_content, - headers=_headers, - params=_params, - ) + _request = build_genai_import_custom_model_request( + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, ) _request.url = self._client.format_url(_request.url) @@ -244148,12 +254188,14 @@ def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=nam return cast(JSON, deserialized) # type: ignore @distributed_trace - def list_evaluation_metrics(self, **kwargs: Any) -> JSON: + def delete_custom_model(self, uuid: str, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Evaluation Metrics. + """Delete Custom Model. - To list all evaluation metrics, send a GET request to ``/v2/gen-ai/evaluation_metrics``. + To delete a custom model, send a DELETE request to ``/v2/genai/custom_models/{uuid}``. + :param uuid: UUID of the custom model to delete. Required. + :type uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -244163,34 +254205,11 @@ def list_evaluation_metrics(self, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "metrics": [ - { - "category": "METRIC_CATEGORY_UNSPECIFIED", # Optional. - Default value is "METRIC_CATEGORY_UNSPECIFIED". Known values are: - "METRIC_CATEGORY_UNSPECIFIED", "METRIC_CATEGORY_CORRECTNESS", - "METRIC_CATEGORY_USER_OUTCOMES", "METRIC_CATEGORY_SAFETY_AND_SECURITY", - "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". - "description": "str", # Optional. - "inverted": bool, # Optional. If true, the metric is - inverted, meaning that a lower value is better. - "is_metric_goal": bool, # Optional. - "metric_name": "str", # Optional. - "metric_rank": 0, # Optional. - "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. - Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". - "metric_uuid": "str", # Optional. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "range_max": 0.0, # Optional. The maximum value for the - metric. - "range_min": 0.0 # Optional. The minimum value for the - metric. - } - ] + "error": "str", # Optional. Error message if deletion failed. + "status": "DELETE_CUSTOM_MODEL_STATUS_UNSPECIFIED" # Optional. Default value + is "DELETE_CUSTOM_MODEL_STATUS_UNSPECIFIED". Status of delete operation. Known + values are: "DELETE_CUSTOM_MODEL_STATUS_UNSPECIFIED", + "DELETE_CUSTOM_MODEL_STATUS_SUCCESS", and "DELETE_CUSTOM_MODEL_STATUS_FAIL". } # response body for status code(s): 404 response == { @@ -244222,7 +254241,8 @@ def list_evaluation_metrics(self, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_evaluation_metrics_request( + _request = build_genai_delete_custom_model_request( + uuid=uuid, headers=_headers, params=_params, ) @@ -244282,18 +254302,22 @@ def list_evaluation_metrics(self, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - def run_evaluation_test_case( + def update_custom_model_metadata( self, + uuid: str, body: Optional[JSON] = None, *, content_type: str = "application/json", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Run an Evaluation Test Case. + """Update Custom Model Metadata. - To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. + To update custom model metadata, send a PATCH request to + ``/v2/gen-ai/custom_models/{uuid}/metadata``. + :param uuid: UUID of the custom model to update. Required. + :type uuid: str :param body: Default value is None. :type body: JSON :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. @@ -244308,23 +254332,105 @@ def run_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_deployment_names": [ - "str" # Optional. Agent deployment names to run the test case - against (ADK agent workspaces). - ], - "agent_uuids": [ - "str" # Optional. Agent UUIDs to run the test case against (legacy - agents). - ], - "run_name": "str", # Optional. The name of the run. - "test_case_uuid": "str" # Optional. Test-case UUID to run. + "description": "str", # Optional. + "name": "str", # Optional. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "uuid": "str" # Optional. UUID of the custom model to update. } # response body for status code(s): 200 response == { - "evaluation_run_uuids": [ - "str" # Optional. - ] + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + } } # response body for status code(s): 404 response == { @@ -244340,18 +254446,22 @@ def run_evaluation_test_case( """ @overload - def run_evaluation_test_case( + def update_custom_model_metadata( self, + uuid: str, body: Optional[IO[bytes]] = None, *, content_type: str = "application/json", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Run an Evaluation Test Case. + """Update Custom Model Metadata. - To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. + To update custom model metadata, send a PATCH request to + ``/v2/gen-ai/custom_models/{uuid}/metadata``. + :param uuid: UUID of the custom model to update. Required. + :type uuid: str :param body: Default value is None. :type body: IO[bytes] :keyword content_type: Body Parameter content-type. Content type parameter for binary body. @@ -244366,9 +254476,93 @@ def run_evaluation_test_case( # response body for status code(s): 200 response == { - "evaluation_run_uuids": [ - "str" # Optional. - ] + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + } } # response body for status code(s): 404 response == { @@ -244384,14 +254578,17 @@ def run_evaluation_test_case( """ @distributed_trace - def run_evaluation_test_case( - self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + def update_custom_model_metadata( + self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Run an Evaluation Test Case. + """Update Custom Model Metadata. - To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. + To update custom model metadata, send a PATCH request to + ``/v2/gen-ai/custom_models/{uuid}/metadata``. + :param uuid: UUID of the custom model to update. Required. + :type uuid: str :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] :return: JSON object @@ -244403,23 +254600,105 @@ def run_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_deployment_names": [ - "str" # Optional. Agent deployment names to run the test case - against (ADK agent workspaces). - ], - "agent_uuids": [ - "str" # Optional. Agent UUIDs to run the test case against (legacy - agents). - ], - "run_name": "str", # Optional. The name of the run. - "test_case_uuid": "str" # Optional. Test-case UUID to run. + "description": "str", # Optional. + "name": "str", # Optional. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "uuid": "str" # Optional. UUID of the custom model to update. } # response body for status code(s): 200 response == { - "evaluation_run_uuids": [ - "str" # Optional. - ] + "model": { + "active_deployments": [ + { + "created_at": "str", # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was created. + "endpoints": { + "private_endpoint_fqdn": "str", # Optional. + Private FQDN for the deployment. + "public_endpoint_fqdn": "str" # Optional. + Public FQDN for the deployment. + }, + "id": "str", # Optional. Unique identifier (UUID) of + the dedicated inference deployment. + "name": "str", # Optional. Human-readable name of + the dedicated inference deployment. + "region_slug": "str", # Optional. Slug of the region + where the dedicated inference deployment is running (e.g. "atl1"). + "state": "str", # Optional. Current lifecycle state + of the dedicated inference deployment (e.g. "ACTIVE", + "PROVISIONING"). + "updated_at": "str" # Optional. RFC 3339 timestamp + indicating when the dedicated inference deployment was last updated. + } + ], + "architecture": "str", # Optional. Model architecture type + (free-form string from config.json). + "config_json": {}, # Optional. Raw config.json contents from the + model repository. + "context_length": 0, # Optional. Maximum context length supported by + the model. + "cost_estimate_per_month": 0, # Optional. Estimated monthly cost in + dollars for hosting. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was created. + "description": "str", # Optional. Description of the custom model. + "file_count": 0, # Optional. Number of files in the model. + "input_modalities": [ + "str" # Optional. Input modalities supported (e.g., text, + image). + ], + "license": "str", # Optional. License under which the model is + distributed. + "name": "str", # Optional. Name of the custom model. + "output_modalities": [ + "str" # Optional. Output modalities supported (e.g., text, + image). + ], + "parameters": "str", # Optional. Number of parameters in the model. + "source_ref": { + "access_type": "ACCESS_TYPE_UNSPECIFIED", # Optional. + Default value is "ACCESS_TYPE_UNSPECIFIED". Access level required for the + model repository. Known values are: "ACCESS_TYPE_UNSPECIFIED", + "ACCESS_TYPE_PUBLIC", "ACCESS_TYPE_PRIVATE", and "ACCESS_TYPE_GATED". + "bucket": "str", # Optional. Spaces bucket name. + "commit_sha": "str", # Optional. Git commit SHA of the model + version. + "hf_token": "str", # Optional. User-provided HuggingFace + token for gated/private models (not persisted in source_ref). + "prefix": "str", # Optional. Object prefix path in the + bucket. + "region": "str", # Optional. Spaces bucket region. + "repo_id": "str" # Optional. Huggingface repository + identifier. + }, + "source_type": "SOURCE_TYPE_UNSPECIFIED", # Optional. Default value + is "SOURCE_TYPE_UNSPECIFIED". Source from which the model was imported. Known + values are: "SOURCE_TYPE_UNSPECIFIED", "SOURCE_TYPE_HUGGINGFACE", + "SOURCE_TYPE_SPACES_BUCKET", "SOURCE_TYPE_SDK_UPLOAD", and + "SOURCE_TYPE_FINE_TUNING". + "status": "STATUS_UNSPECIFIED", # Optional. Default value is + "STATUS_UNSPECIFIED". Import and deployment status of the custom model. Known + values are: "STATUS_UNSPECIFIED", "STATUS_IMPORTING", "STATUS_READY", + "STATUS_FAILED", and "STATUS_DELETED". + "storage_region": "str", # Optional. Region of the Spaces bucket + where model files are stored. + "tags": { + "tags": [ + "str" # Optional. List of tag strings. + ] + }, + "team_id": "str", # Optional. Team that owns the model. + "total_size_bytes": "str", # Optional. Total size of model files in + bytes. + "updated_at": "2020-02-20 00:00:00", # Optional. Timestamp when the + model was last updated. + "uuid": "str" # Optional. Unique identifier for the custom model. + } } # response body for status code(s): 404 response == { @@ -244465,7 +254744,8 @@ def run_evaluation_test_case( else: _json = None - _request = build_genai_run_evaluation_test_case_request( + _request = build_genai_update_custom_model_metadata_request( + uuid=uuid, content_type=content_type, json=_json, content=_content, @@ -244527,16 +254807,116 @@ def run_evaluation_test_case( return cast(JSON, deserialized) # type: ignore + @overload + def create_evaluation_dataset( + self, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Create Evaluation Dataset. + + To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. + + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and "EVALUATION_DATASET_TYPE_MODEL". + "file_upload_dataset": { + "original_file_name": "str", # Optional. The original file name. + "size_in_bytes": "str", # Optional. The size of the file in bytes. + "stored_object_key": "str" # Optional. The object key the file was + stored as. + }, + "name": "str" # Optional. The name of the agent evaluation dataset. + } + + # response body for status code(s): 200 + response == { + "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + def create_evaluation_dataset( + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Create Evaluation Dataset. + + To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + @distributed_trace - def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> JSON: + def create_evaluation_dataset( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: # pylint: disable=line-too-long - """Retrieve Information About an Existing Evaluation Run. + """Create Evaluation Dataset. - To retrive information about an existing evaluation run, send a GET request to - ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}``. + To create an evaluation dataset, send a POST request to ``/v2/gen-ai/evaluation_datasets``. - :param evaluation_run_uuid: Evaluation run UUID. Required. - :type evaluation_run_uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -244544,73 +254924,24 @@ def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> JSON: Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and "EVALUATION_DATASET_TYPE_MODEL". + "file_upload_dataset": { + "original_file_name": "str", # Optional. The original file name. + "size_in_bytes": "str", # Optional. The size of the file in bytes. + "stored_object_key": "str" # Optional. The object key the file was + stored as. + }, + "name": "str" # Optional. The name of the agent evaluation dataset. + } + # response body for status code(s): 200 response == { - "evaluation_run": { - "agent_deleted": bool, # Optional. Whether agent is deleted. - "agent_deployment_name": "str", # Optional. The agent deployment - name. - "agent_name": "str", # Optional. Agent name. - "agent_uuid": "str", # Optional. Agent UUID. - "agent_version_hash": "str", # Optional. Version hash. - "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. - "created_by_user_email": "str", # Optional. - "created_by_user_id": "str", # Optional. - "error_description": "str", # Optional. The error description. - "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. - "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation - test case workspace uuid. - "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. - "pass_status": bool, # Optional. The pass status of the evaluation - run based on the star metric. - "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. - "run_level_metric_results": [ - { - "error_description": "str", # Optional. Error - description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", - # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known - values are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the - metric as a number. - "reasoning": "str", # Optional. Reasoning of the - metric result. - "string_value": "str" # Optional. The value of the - metric as a string. - } - ], - "run_name": "str", # Optional. Run name. - "star_metric_result": { - "error_description": "str", # Optional. Error description if - the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the metric as - a number. - "reasoning": "str", # Optional. Reasoning of the metric - result. - "string_value": "str" # Optional. The value of the metric as - a string. - }, - "started_at": "2020-02-20 00:00:00", # Optional. Run start time. - "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default - value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known - values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", - "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", - "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", - "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and - "EVALUATION_RUN_FAILED". - "test_case_description": "str", # Optional. Test case description. - "test_case_name": "str", # Optional. Test case name. - "test_case_uuid": "str", # Optional. Test-case UUID. - "test_case_version": 0 # Optional. Test-case-version. - } + "evaluation_dataset_uuid": "str" # Optional. Evaluation dataset uuid. } # response body for status code(s): 404 response == { @@ -244637,13 +254968,29 @@ def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> JSON: } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = kwargs.pop("headers", {}) or {} + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_run_request( - evaluation_run_uuid=evaluation_run_uuid, + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_create_evaluation_dataset_request( + content_type=content_type, + json=_json, + content=_content, headers=_headers, params=_params, ) @@ -244702,27 +255049,25 @@ def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore - @distributed_trace - def get_evaluation_run_results( + @overload + def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long self, - evaluation_run_uuid: str, + body: Optional[JSON] = None, *, - page: Optional[int] = None, - per_page: Optional[int] = None, + content_type: str = "application/json", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Retrieve Results of an Evaluation Run. + """Create Presigned URLs for Evaluation Dataset File Upload. - To retrieve results of an evaluation run, send a GET request to - ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results``. + To create presigned URLs for evaluation dataset file upload, send a POST request to + ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. - :param evaluation_run_uuid: Evaluation run UUID. Required. - :type evaluation_run_uuid: str - :keyword page: Page number. Default value is None. - :paramtype page: int - :keyword per_page: Items per page. Default value is None. - :paramtype per_page: int + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -244730,184 +255075,147 @@ def get_evaluation_run_results( Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + # response body for status code(s): 200 response == { - "evaluation_run": { - "agent_deleted": bool, # Optional. Whether agent is deleted. - "agent_deployment_name": "str", # Optional. The agent deployment - name. - "agent_name": "str", # Optional. Agent name. - "agent_uuid": "str", # Optional. Agent UUID. - "agent_version_hash": "str", # Optional. Version hash. - "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. - "created_by_user_email": "str", # Optional. - "created_by_user_id": "str", # Optional. - "error_description": "str", # Optional. The error description. - "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. - "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation - test case workspace uuid. - "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. - "pass_status": bool, # Optional. The pass status of the evaluation - run based on the star metric. - "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. - "run_level_metric_results": [ - { - "error_description": "str", # Optional. Error - description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", - # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known - values are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the - metric as a number. - "reasoning": "str", # Optional. Reasoning of the - metric result. - "string_value": "str" # Optional. The value of the - metric as a string. - } - ], - "run_name": "str", # Optional. Run name. - "star_metric_result": { - "error_description": "str", # Optional. Error description if - the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the metric as - a number. - "reasoning": "str", # Optional. Reasoning of the metric - result. - "string_value": "str" # Optional. The value of the metric as - a string. - }, - "started_at": "2020-02-20 00:00:00", # Optional. Run start time. - "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default - value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known - values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", - "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", - "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", - "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and - "EVALUATION_RUN_FAILED". - "test_case_description": "str", # Optional. Test case description. - "test_case_name": "str", # Optional. Test case name. - "test_case_uuid": "str", # Optional. Test-case UUID. - "test_case_version": 0 # Optional. Test-case-version. - }, - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. - "total": 0 # Optional. Total amount of items over all pages. - }, - "prompts": [ + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Evaluation Dataset File Upload. + + To create presigned URLs for evaluation dataset file upload, send a POST request to + ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ { - "evaluation_trace_spans": [ - { - "created_at": "2020-02-20 00:00:00", # - Optional. When the span was created. - "input": {}, # Optional. Input data for the - span (flexible structure - can be messages array, string, etc.). - "name": "str", # Optional. Name/identifier - for the span. - "output": {}, # Optional. Output data from - the span (flexible structure - can be message, string, etc.). - "retriever_chunks": [ - { - "chunk_usage_pct": 0.0, # - Optional. The usage percentage of the chunk. - "chunk_used": bool, # - Optional. Indicates if the chunk was used in the prompt. - "index_uuid": "str", # - Optional. The index uuid (Knowledge Base) of the chunk. - "source_name": "str", # - Optional. The source name for the chunk, e.g., the file - name or document title. - "text": "str" # Optional. - Text content of the chunk. - } - ], - "span_level_metric_results": [ - { - "error_description": "str", - # Optional. Error description if the metric could not be - calculated. - "metric_name": "str", # - Optional. Metric name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default - value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values - are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", - and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # - Optional. The value of the metric as a number. - "reasoning": "str", # - Optional. Reasoning of the metric result. - "string_value": "str" # - Optional. The value of the metric as a string. - } - ], - "type": "TRACE_SPAN_TYPE_UNKNOWN" # - Optional. Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of - spans in a trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", - "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", and - "TRACE_SPAN_TYPE_TOOL". - } - ], - "ground_truth": "str", # Optional. The ground truth for the - prompt. - "input": "str", # Optional. The prompt level results. - "input_tokens": "str", # Optional. The number of input - tokens used in the prompt. - "output": "str", # Optional. The prompt level results. - "output_tokens": "str", # Optional. The number of output - tokens used in the prompt. - "prompt_chunks": [ - { - "chunk_usage_pct": 0.0, # Optional. The - usage percentage of the chunk. - "chunk_used": bool, # Optional. Indicates if - the chunk was used in the prompt. - "index_uuid": "str", # Optional. The index - uuid (Knowledge Base) of the chunk. - "source_name": "str", # Optional. The source - name for the chunk, e.g., the file name or document title. - "text": "str" # Optional. Text content of - the chunk. - } - ], - "prompt_id": 0, # Optional. Prompt ID. - "prompt_level_metric_results": [ - { - "error_description": "str", # Optional. - Error description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric - name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is - "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value - of the metric as a number. - "reasoning": "str", # Optional. Reasoning of - the metric result. - "string_value": "str" # Optional. The value - of the metric as a string. - } - ], - "trace_id": "str" # Optional. The trace id for the prompt. + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace + def create_evaluation_dataset_file_upload_presigned_urls( # pylint: disable=name-too-long + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Evaluation Dataset File Upload. + + To create presigned URLs for evaluation dataset file upload, send a POST request to + ``/v2/gen-ai/evaluation_datasets/file_upload_presigned_urls``. + + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. } ] } @@ -244936,17 +255244,33 @@ def get_evaluation_run_results( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = kwargs.pop("headers", {}) or {} + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_run_results_request( - evaluation_run_uuid=evaluation_run_uuid, - page=page, - per_page=per_page, - headers=_headers, - params=_params, + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = ( + build_genai_create_evaluation_dataset_file_upload_presigned_urls_request( + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) ) _request.url = self._client.format_url(_request.url) @@ -245004,19 +255328,17 @@ def get_evaluation_run_results( return cast(JSON, deserialized) # type: ignore @distributed_trace - def get_evaluation_run_prompt_results( - self, evaluation_run_uuid: str, prompt_id: int, **kwargs: Any + def get_evaluation_dataset_download_url( + self, dataset_uuid: str, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Retrieve Results of an Evaluation Run Prompt. + """Get Download URL for Evaluation Dataset. - To retrieve results of an evaluation run, send a GET request to - ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results/{prompt_id}``. + To get a presigned download URL for an evaluation dataset, send a GET request to + ``/v2/genai/evaluation_datasets/{dataset_uuid}/download_url``. - :param evaluation_run_uuid: Evaluation run UUID. Required. - :type evaluation_run_uuid: str - :param prompt_id: Prompt ID to get results for. Required. - :type prompt_id: int + :param dataset_uuid: UUID of the evaluation dataset. Required. + :type dataset_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -245026,101 +255348,9 @@ def get_evaluation_run_prompt_results( # response body for status code(s): 200 response == { - "prompt": { - "evaluation_trace_spans": [ - { - "created_at": "2020-02-20 00:00:00", # Optional. - When the span was created. - "input": {}, # Optional. Input data for the span - (flexible structure - can be messages array, string, etc.). - "name": "str", # Optional. Name/identifier for the - span. - "output": {}, # Optional. Output data from the span - (flexible structure - can be message, string, etc.). - "retriever_chunks": [ - { - "chunk_usage_pct": 0.0, # Optional. - The usage percentage of the chunk. - "chunk_used": bool, # Optional. - Indicates if the chunk was used in the prompt. - "index_uuid": "str", # Optional. The - index uuid (Knowledge Base) of the chunk. - "source_name": "str", # Optional. - The source name for the chunk, e.g., the file name or - document title. - "text": "str" # Optional. Text - content of the chunk. - } - ], - "span_level_metric_results": [ - { - "error_description": "str", # - Optional. Error description if the metric could not be - calculated. - "metric_name": "str", # Optional. - Metric name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value - is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The - value of the metric as a number. - "reasoning": "str", # Optional. - Reasoning of the metric result. - "string_value": "str" # Optional. - The value of the metric as a string. - } - ], - "type": "TRACE_SPAN_TYPE_UNKNOWN" # Optional. - Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of spans in a - trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", - "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", and - "TRACE_SPAN_TYPE_TOOL". - } - ], - "ground_truth": "str", # Optional. The ground truth for the prompt. - "input": "str", # Optional. - "input_tokens": "str", # Optional. The number of input tokens used - in the prompt. - "output": "str", # Optional. - "output_tokens": "str", # Optional. The number of output tokens used - in the prompt. - "prompt_chunks": [ - { - "chunk_usage_pct": 0.0, # Optional. The usage - percentage of the chunk. - "chunk_used": bool, # Optional. Indicates if the - chunk was used in the prompt. - "index_uuid": "str", # Optional. The index uuid - (Knowledge Base) of the chunk. - "source_name": "str", # Optional. The source name - for the chunk, e.g., the file name or document title. - "text": "str" # Optional. Text content of the chunk. - } - ], - "prompt_id": 0, # Optional. Prompt ID. - "prompt_level_metric_results": [ - { - "error_description": "str", # Optional. Error - description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric name. - "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", - # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known - values are: "METRIC_VALUE_TYPE_UNSPECIFIED", - "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and - "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value of the - metric as a number. - "reasoning": "str", # Optional. Reasoning of the - metric result. - "string_value": "str" # Optional. The value of the - metric as a string. - } - ], - "trace_id": "str" # Optional. The trace id for the prompt. - } + "download_url": "str", # Optional. The presigned URL to download the dataset + file. + "expires_at": "2020-02-20 00:00:00" # Optional. The time the URL expires at. } # response body for status code(s): 404 response == { @@ -245152,9 +255382,8 @@ def get_evaluation_run_prompt_results( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_run_prompt_results_request( - evaluation_run_uuid=evaluation_run_uuid, - prompt_id=prompt_id, + _request = build_genai_get_evaluation_dataset_download_url_request( + dataset_uuid=dataset_uuid, headers=_headers, params=_params, ) @@ -245214,11 +255443,11 @@ def get_evaluation_run_prompt_results( return cast(JSON, deserialized) # type: ignore @distributed_trace - def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: + def list_evaluation_metrics(self, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Evaluation Test Cases. + """List Evaluation Metrics. - To list all evaluation test cases, send a GET request to ``/v2/gen-ai/evaluation_test_cases``. + To list all evaluation metrics, send a GET request to ``/v2/gen-ai/evaluation_metrics``. :return: JSON object :rtype: JSON @@ -245229,120 +255458,40 @@ def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "evaluation_test_cases": [ + "metrics": [ { - "archived_at": "2020-02-20 00:00:00", # Optional. - Alternative way of authentication for internal usage only - should not be - exposed to public api. - "created_at": "2020-02-20 00:00:00", # Optional. Alternative - way of authentication for internal usage only - should not be exposed to - public api. - "created_by_user_email": "str", # Optional. Alternative way - of authentication for internal usage only - should not be exposed to - public api. - "created_by_user_id": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "dataset": { - "created_at": "2020-02-20 00:00:00", # Optional. - Time created at. - "dataset_name": "str", # Optional. Name of the - dataset. - "dataset_uuid": "str", # Optional. UUID of the - dataset. - "file_size": "str", # Optional. The size of the - dataset uploaded file in bytes. - "has_ground_truth": bool, # Optional. Does the - dataset have a ground truth column?. - "row_count": 0 # Optional. Number of rows in the - dataset. - }, - "dataset_name": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "dataset_uuid": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "description": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "latest_version_number_of_runs": 0, # Optional. Alternative - way of authentication for internal usage only - should not be exposed to - public api. - "metrics": [ - { - "category": "METRIC_CATEGORY_UNSPECIFIED", # - Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known - values are: "METRIC_CATEGORY_UNSPECIFIED", - "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", - "METRIC_CATEGORY_SAFETY_AND_SECURITY", - "METRIC_CATEGORY_CONTEXT_QUALITY", and - "METRIC_CATEGORY_MODEL_FIT". - "description": "str", # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "inverted": bool, # Optional. If true, the - metric is inverted, meaning that a lower value is better. - "is_metric_goal": bool, # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "metric_name": "str", # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "metric_rank": 0, # Optional. Alternative - way of authentication for internal usage only - should not be - exposed to public api. - "metric_type": "METRIC_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known - values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and "METRIC_TYPE_RAG_AND_TOOL". - "metric_uuid": "str", # Optional. - Alternative way of authentication for internal usage only - - should not be exposed to public api. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is - "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "range_max": 0.0, # Optional. The maximum - value for the metric. - "range_min": 0.0 # Optional. The minimum - value for the metric. - } - ], - "name": "str", # Optional. Alternative way of authentication - for internal usage only - should not be exposed to public api. - "star_metric": { - "metric_uuid": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to - public api. - "name": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to - public api. - "success_threshold": 0.0, # Optional. The success - threshold for the star metric. This is a value that the metric must - reach to be considered successful. - "success_threshold_pct": 0 # Optional. The success - threshold for the star metric. This is a percentage value between 0 - and 100. - }, - "test_case_uuid": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "total_runs": 0, # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "updated_at": "2020-02-20 00:00:00", # Optional. Alternative - way of authentication for internal usage only - should not be exposed to - public api. - "updated_by_user_email": "str", # Optional. Alternative way - of authentication for internal usage only - should not be exposed to - public api. - "updated_by_user_id": "str", # Optional. Alternative way of - authentication for internal usage only - should not be exposed to public - api. - "version": 0 # Optional. Alternative way of authentication - for internal usage only - should not be exposed to public api. + "category": "METRIC_CATEGORY_UNSPECIFIED", # Optional. + Default value is "METRIC_CATEGORY_UNSPECIFIED". Known values are: + "METRIC_CATEGORY_UNSPECIFIED", "METRIC_CATEGORY_CORRECTNESS", + "METRIC_CATEGORY_USER_OUTCOMES", "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", # + Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. + "metric_name": "str", # Optional. + "metric_rank": 0, # Optional. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. + Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for the + metric. + "range_min": 0.0 # Optional. The minimum value for the + metric. } ] } @@ -245376,7 +255525,7 @@ def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_evaluation_test_cases_request( + _request = build_genai_list_evaluation_metrics_request( headers=_headers, params=_params, ) @@ -245436,7 +255585,7 @@ def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - def create_evaluation_test_case( + def run_evaluation_test_case( self, body: Optional[JSON] = None, *, @@ -245444,9 +255593,9 @@ def create_evaluation_test_case( **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Test Case. + """Run an Evaluation Test Case. - To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. + To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. :param body: Default value is None. :type body: JSON @@ -245462,29 +255611,23 @@ def create_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_workspace_name": "str", # Optional. - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": [ - "str" # Optional. Full metric list to use for evaluation test case. + "agent_deployment_names": [ + "str" # Optional. Agent deployment names to run the test case + against (ADK agent workspaces). ], - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "workspace_uuid": "str" # Optional. The workspace uuid. + "agent_uuids": [ + "str" # Optional. Agent UUIDs to run the test case against (legacy + agents). + ], + "run_name": "str", # Optional. The name of the run. + "test_case_uuid": "str" # Optional. Test-case UUID to run. } # response body for status code(s): 200 response == { - "test_case_uuid": "str" # Optional. Test"u2011case UUID. + "evaluation_run_uuids": [ + "str" # Optional. + ] } # response body for status code(s): 404 response == { @@ -245500,7 +255643,7 @@ def create_evaluation_test_case( """ @overload - def create_evaluation_test_case( + def run_evaluation_test_case( self, body: Optional[IO[bytes]] = None, *, @@ -245508,9 +255651,9 @@ def create_evaluation_test_case( **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Test Case. + """Run an Evaluation Test Case. - To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. + To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. :param body: Default value is None. :type body: IO[bytes] @@ -245526,7 +255669,9 @@ def create_evaluation_test_case( # response body for status code(s): 200 response == { - "test_case_uuid": "str" # Optional. Test"u2011case UUID. + "evaluation_run_uuids": [ + "str" # Optional. + ] } # response body for status code(s): 404 response == { @@ -245542,13 +255687,13 @@ def create_evaluation_test_case( """ @distributed_trace - def create_evaluation_test_case( + def run_evaluation_test_case( self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Create Evaluation Test Case. + """Run an Evaluation Test Case. - To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. + To run an evaluation test case, send a POST request to ``/v2/gen-ai/evaluation_runs``. :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] @@ -245561,29 +255706,23 @@ def create_evaluation_test_case( # JSON input template you can fill out and use as your body input. body = { - "agent_workspace_name": "str", # Optional. - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": [ - "str" # Optional. Full metric list to use for evaluation test case. + "agent_deployment_names": [ + "str" # Optional. Agent deployment names to run the test case + against (ADK agent workspaces). ], - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "workspace_uuid": "str" # Optional. The workspace uuid. + "agent_uuids": [ + "str" # Optional. Agent UUIDs to run the test case against (legacy + agents). + ], + "run_name": "str", # Optional. The name of the run. + "test_case_uuid": "str" # Optional. Test-case UUID to run. } # response body for status code(s): 200 response == { - "test_case_uuid": "str" # Optional. Test"u2011case UUID. + "evaluation_run_uuids": [ + "str" # Optional. + ] } # response body for status code(s): 404 response == { @@ -245629,7 +255768,7 @@ def create_evaluation_test_case( else: _json = None - _request = build_genai_create_evaluation_test_case_request( + _request = build_genai_run_evaluation_test_case_request( content_type=content_type, json=_json, content=_content, @@ -245692,23 +255831,15 @@ def create_evaluation_test_case( return cast(JSON, deserialized) # type: ignore @distributed_trace - def list_evaluation_runs_by_test_case( - self, - evaluation_test_case_uuid: str, - *, - evaluation_test_case_version: Optional[int] = None, - **kwargs: Any, - ) -> JSON: + def get_evaluation_run(self, evaluation_run_uuid: str, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Evaluation Runs by Test Case. + """Retrieve Information About an Existing Evaluation Run. - To list all evaluation runs by test case, send a GET request to - ``/v2/gen-ai/evaluation_test_cases/{evaluation_test_case_uuid}/evaluation_runs``. + To retrive information about an existing evaluation run, send a GET request to + ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}``. - :param evaluation_test_case_uuid: Evaluation run UUID. Required. - :type evaluation_test_case_uuid: str - :keyword evaluation_test_case_version: Version of the test case. Default value is None. - :paramtype evaluation_test_case_version: int + :param evaluation_run_uuid: Evaluation run UUID. Required. + :type evaluation_run_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -245718,53 +255849,26 @@ def list_evaluation_runs_by_test_case( # response body for status code(s): 200 response == { - "evaluation_runs": [ - { - "agent_deleted": bool, # Optional. Whether agent is deleted. - "agent_deployment_name": "str", # Optional. The agent - deployment name. - "agent_name": "str", # Optional. Agent name. - "agent_uuid": "str", # Optional. Agent UUID. - "agent_version_hash": "str", # Optional. Version hash. - "agent_workspace_uuid": "str", # Optional. Agent workspace - uuid. - "created_by_user_email": "str", # Optional. List of - evaluation runs. - "created_by_user_id": "str", # Optional. List of evaluation - runs. - "error_description": "str", # Optional. The error - description. - "evaluation_run_uuid": "str", # Optional. Evaluation run - UUID. - "evaluation_test_case_workspace_uuid": "str", # Optional. - Evaluation test case workspace uuid. - "finished_at": "2020-02-20 00:00:00", # Optional. Run end - time. - "pass_status": bool, # Optional. The pass status of the - evaluation run based on the star metric. - "queued_at": "2020-02-20 00:00:00", # Optional. Run queued - time. - "run_level_metric_results": [ - { - "error_description": "str", # Optional. - Error description if the metric could not be calculated. - "metric_name": "str", # Optional. Metric - name. - "metric_value_type": - "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is - "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: - "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", - "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "number_value": 0.0, # Optional. The value - of the metric as a number. - "reasoning": "str", # Optional. Reasoning of - the metric result. - "string_value": "str" # Optional. The value - of the metric as a string. - } - ], - "run_name": "str", # Optional. Run name. - "star_metric_result": { + "evaluation_run": { + "agent_deleted": bool, # Optional. Whether agent is deleted. + "agent_deployment_name": "str", # Optional. The agent deployment + name. + "agent_name": "str", # Optional. Agent name. + "agent_uuid": "str", # Optional. Agent UUID. + "agent_version_hash": "str", # Optional. Version hash. + "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. + "created_by_user_email": "str", # Optional. + "created_by_user_id": "str", # Optional. + "error_description": "str", # Optional. The error description. + "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. + "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation + test case workspace uuid. + "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. + "pass_status": bool, # Optional. The pass status of the evaluation + run based on the star metric. + "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. + "run_level_metric_results": [ + { "error_description": "str", # Optional. Error description if the metric could not be calculated. "metric_name": "str", # Optional. Metric name. @@ -245779,33 +255883,47 @@ def list_evaluation_runs_by_test_case( metric result. "string_value": "str" # Optional. The value of the metric as a string. - }, - "started_at": "2020-02-20 00:00:00", # Optional. Run start - time. - "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. - Default value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run - Statuses. Known values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", - "EVALUATION_RUN_QUEUED", "EVALUATION_RUN_RUNNING_DATASET", - "EVALUATION_RUN_EVALUATING_RESULTS", "EVALUATION_RUN_CANCELLING", - "EVALUATION_RUN_CANCELLED", "EVALUATION_RUN_SUCCESSFUL", - "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and "EVALUATION_RUN_FAILED". - "test_case_description": "str", # Optional. Test case - description. - "test_case_name": "str", # Optional. Test case name. - "test_case_uuid": "str", # Optional. Test-case UUID. - "test_case_version": 0 # Optional. Test-case-version. - } - ] - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support + } + ], + "run_name": "str", # Optional. Run name. + "star_metric_result": { + "error_description": "str", # Optional. Error description if + the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the metric as + a number. + "reasoning": "str", # Optional. Reasoning of the metric + result. + "string_value": "str" # Optional. The value of the metric as + a string. + }, + "started_at": "2020-02-20 00:00:00", # Optional. Run start time. + "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default + value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known + values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", + "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", + "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", + "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "EVALUATION_RUN_FAILED". + "test_case_description": "str", # Optional. Test case description. + "test_case_name": "str", # Optional. Test case name. + "test_case_uuid": "str", # Optional. Test-case UUID. + "test_case_version": 0 # Optional. Test-case-version. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support tickets to help identify the issue. } """ @@ -245827,9 +255945,8 @@ def list_evaluation_runs_by_test_case( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_evaluation_runs_by_test_case_request( - evaluation_test_case_uuid=evaluation_test_case_uuid, - evaluation_test_case_version=evaluation_test_case_version, + _request = build_genai_get_evaluation_run_request( + evaluation_run_uuid=evaluation_run_uuid, headers=_headers, params=_params, ) @@ -245889,23 +256006,26 @@ def list_evaluation_runs_by_test_case( return cast(JSON, deserialized) # type: ignore @distributed_trace - def get_evaluation_test_case( + def get_evaluation_run_results( self, - test_case_uuid: str, + evaluation_run_uuid: str, *, - evaluation_test_case_version: Optional[int] = None, + page: Optional[int] = None, + per_page: Optional[int] = None, **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Retrieve Information About an Existing Evaluation Test Case. + """Retrieve Results of an Evaluation Run. - To retrive information about an existing evaluation test case, send a GET request to - ``/v2/gen-ai/evaluation_test_case/{test_case_uuid}``. + To retrieve results of an evaluation run, send a GET request to + ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results``. - :param test_case_uuid: The test case uuid to retrieve. Required. - :type test_case_uuid: str - :keyword evaluation_test_case_version: Version of the test case. Default value is None. - :paramtype evaluation_test_case_version: int + :param evaluation_run_uuid: Evaluation run UUID. Required. + :type evaluation_run_uuid: str + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -245915,73 +256035,346 @@ def get_evaluation_test_case( # response body for status code(s): 200 response == { - "evaluation_test_case": { - "archived_at": "2020-02-20 00:00:00", # Optional. - "created_at": "2020-02-20 00:00:00", # Optional. + "evaluation_run": { + "agent_deleted": bool, # Optional. Whether agent is deleted. + "agent_deployment_name": "str", # Optional. The agent deployment + name. + "agent_name": "str", # Optional. Agent name. + "agent_uuid": "str", # Optional. Agent UUID. + "agent_version_hash": "str", # Optional. Version hash. + "agent_workspace_uuid": "str", # Optional. Agent workspace uuid. "created_by_user_email": "str", # Optional. "created_by_user_id": "str", # Optional. - "dataset": { - "created_at": "2020-02-20 00:00:00", # Optional. Time - created at. - "dataset_name": "str", # Optional. Name of the dataset. - "dataset_uuid": "str", # Optional. UUID of the dataset. - "file_size": "str", # Optional. The size of the dataset - uploaded file in bytes. - "has_ground_truth": bool, # Optional. Does the dataset have - a ground truth column?. - "row_count": 0 # Optional. Number of rows in the dataset. - }, - "dataset_name": "str", # Optional. - "dataset_uuid": "str", # Optional. - "description": "str", # Optional. - "latest_version_number_of_runs": 0, # Optional. - "metrics": [ + "error_description": "str", # Optional. The error description. + "evaluation_run_uuid": "str", # Optional. Evaluation run UUID. + "evaluation_test_case_workspace_uuid": "str", # Optional. Evaluation + test case workspace uuid. + "finished_at": "2020-02-20 00:00:00", # Optional. Run end time. + "pass_status": bool, # Optional. The pass status of the evaluation + run based on the star metric. + "queued_at": "2020-02-20 00:00:00", # Optional. Run queued time. + "run_level_metric_results": [ { - "category": "METRIC_CATEGORY_UNSPECIFIED", # - Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known - values are: "METRIC_CATEGORY_UNSPECIFIED", - "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", - "METRIC_CATEGORY_SAFETY_AND_SECURITY", - "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". - "description": "str", # Optional. - "inverted": bool, # Optional. If true, the metric is - inverted, meaning that a lower value is better. - "is_metric_goal": bool, # Optional. - "metric_name": "str", # Optional. - "metric_rank": 0, # Optional. - "metric_type": "METRIC_TYPE_UNSPECIFIED", # - Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values - are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". - "metric_uuid": "str", # Optional. + "error_description": "str", # Optional. Error + description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". - "range_max": 0.0, # Optional. The maximum value for - the metric. - "range_min": 0.0 # Optional. The minimum value for - the metric. + "number_value": 0.0, # Optional. The value of the + metric as a number. + "reasoning": "str", # Optional. Reasoning of the + metric result. + "string_value": "str" # Optional. The value of the + metric as a string. } ], - "name": "str", # Optional. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold - for the star metric. This is a value that the metric must reach to be - considered successful. - "success_threshold_pct": 0 # Optional. The success threshold - for the star metric. This is a percentage value between 0 and 100. + "run_name": "str", # Optional. Run name. + "star_metric_result": { + "error_description": "str", # Optional. Error description if + the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the metric as + a number. + "reasoning": "str", # Optional. Reasoning of the metric + result. + "string_value": "str" # Optional. The value of the metric as + a string. }, - "test_case_uuid": "str", # Optional. - "total_runs": 0, # Optional. - "updated_at": "2020-02-20 00:00:00", # Optional. - "updated_by_user_email": "str", # Optional. - "updated_by_user_id": "str", # Optional. - "version": 0 # Optional. - } + "started_at": "2020-02-20 00:00:00", # Optional. Run start time. + "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. Default + value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run Statuses. Known + values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", "EVALUATION_RUN_QUEUED", + "EVALUATION_RUN_RUNNING_DATASET", "EVALUATION_RUN_EVALUATING_RESULTS", + "EVALUATION_RUN_CANCELLING", "EVALUATION_RUN_CANCELLED", + "EVALUATION_RUN_SUCCESSFUL", "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "EVALUATION_RUN_FAILED". + "test_case_description": "str", # Optional. Test case description. + "test_case_name": "str", # Optional. Test case name. + "test_case_uuid": "str", # Optional. Test-case UUID. + "test_case_version": 0 # Optional. Test-case-version. + }, + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "prompts": [ + { + "evaluation_trace_spans": [ + { + "created_at": "2020-02-20 00:00:00", # + Optional. When the span was created. + "input": {}, # Optional. Input data for the + span (flexible structure - can be messages array, string, etc.). + "name": "str", # Optional. Name/identifier + for the span. + "output": {}, # Optional. Output data from + the span (flexible structure - can be message, string, etc.). + "retriever_chunks": [ + { + "chunk_usage_pct": 0.0, # + Optional. The usage percentage of the chunk. + "chunk_used": bool, # + Optional. Indicates if the chunk was used in the prompt. + "index_uuid": "str", # + Optional. The index uuid (Knowledge Base) of the chunk. + "source_name": "str", # + Optional. The source name for the chunk, e.g., the file + name or document title. + "text": "str" # Optional. + Text content of the chunk. + } + ], + "span_level_metric_results": [ + { + "error_description": "str", + # Optional. Error description if the metric could not be + calculated. + "metric_name": "str", # + Optional. Metric name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default + value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", + and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # + Optional. The value of the metric as a number. + "reasoning": "str", # + Optional. Reasoning of the metric result. + "string_value": "str" # + Optional. The value of the metric as a string. + } + ], + "spans": [ + { + "agent": { + "agent_type": + "AGENT_TYPE_UNSPECIFIED", # Optional. Default value + is "AGENT_TYPE_UNSPECIFIED". Agent span. Known values + are: "AGENT_TYPE_UNSPECIFIED", "AGENT_TYPE_DEFAULT", + "AGENT_TYPE_PLANNER", "AGENT_TYPE_REACT", + "AGENT_TYPE_REFLECTION", "AGENT_TYPE_ROUTER", + "AGENT_TYPE_CLASSIFIER", "AGENT_TYPE_SUPERVISOR", and + "AGENT_TYPE_JUDGE". + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "redacted_input": + "str", # Optional. Child spans - must contain + between 1 and 999 spans Allowed types: agent, llm, + tool, retriever (not workflow). + "redacted_output": + "str", # Optional. Child spans - must contain + between 1 and 999 spans Allowed types: agent, llm, + tool, retriever (not workflow). + "spans": [ + ... + ] + }, + "created_at": "2020-02-20 + 00:00:00", # Optional. When the span was created. + "input": {}, # Optional. + Input data for the span (flexible structure - can be + messages array, string, etc.). + "llm": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "model": "str", # + Optional. LLM span. + "num_input_tokens": + 0, # Optional. LLM span. + "num_output_tokens": + 0, # Optional. LLM span. + "temperature": 0.0, + # Optional. LLM span. + "time_to_first_token_ns": "str", # Optional. LLM + span. + "tools": [ + {} # + Optional. Tool definitions passed to the model. + ], + "total_tokens": 0 # + Optional. LLM span. + }, + "name": "str", # Optional. + Name/identifier for the span. + "output": {}, # Optional. + Output data from the span (flexible structure - can be + message, string, etc.). + "retriever": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + } + }, + "tool": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "tool_call_id": "str" + # Optional. Tool span. + }, + "type": + "TRACE_SPAN_TYPE_UNKNOWN", # Optional. Default value is + "TRACE_SPAN_TYPE_UNKNOWN". Types of spans in a trace. + Known values are: "TRACE_SPAN_TYPE_UNKNOWN", + "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", + "TRACE_SPAN_TYPE_TOOL", "TRACE_SPAN_TYPE_AGENT", and + "TRACE_SPAN_TYPE_WORKFLOW". + "workflow": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common + optional fields shared by all span types. + "duration_ns": "str", # Optional. Common + optional fields shared by all span types. + "metadata": { + "str": "str" # Optional. Arbitrary + structured metadata. + }, + "status_code": 0, # Optional. Common optional + fields shared by all span types. + "tags": [ + "str" + # Optional. Free-form tags for + filtering/grouping. + ] + }, + "spans": [ + ... + ] + } + } + ], + "type": "TRACE_SPAN_TYPE_UNKNOWN" # + Optional. Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of + spans in a trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", + "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", + "TRACE_SPAN_TYPE_TOOL", "TRACE_SPAN_TYPE_AGENT", and + "TRACE_SPAN_TYPE_WORKFLOW". + } + ], + "ground_truth": "str", # Optional. The ground truth for the + prompt. + "input": "str", # Optional. The prompt level results. + "input_tokens": "str", # Optional. The number of input + tokens used in the prompt. + "output": "str", # Optional. The prompt level results. + "output_tokens": "str", # Optional. The number of output + tokens used in the prompt. + "prompt_chunks": [ + { + "chunk_usage_pct": 0.0, # Optional. The + usage percentage of the chunk. + "chunk_used": bool, # Optional. Indicates if + the chunk was used in the prompt. + "index_uuid": "str", # Optional. The index + uuid (Knowledge Base) of the chunk. + "source_name": "str", # Optional. The source + name for the chunk, e.g., the file name or document title. + "text": "str" # Optional. Text content of + the chunk. + } + ], + "prompt_id": 0, # Optional. Prompt ID. + "prompt_level_metric_results": [ + { + "error_description": "str", # Optional. + Error description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric + name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value + of the metric as a number. + "reasoning": "str", # Optional. Reasoning of + the metric result. + "string_value": "str" # Optional. The value + of the metric as a string. + } + ], + "trace_id": "str" # Optional. The trace id for the prompt. + } + ] } # response body for status code(s): 404 response == { @@ -246013,9 +256406,10 @@ def get_evaluation_test_case( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_evaluation_test_case_request( - test_case_uuid=test_case_uuid, - evaluation_test_case_version=evaluation_test_case_version, + _request = build_genai_get_evaluation_run_results_request( + evaluation_run_uuid=evaluation_run_uuid, + page=page, + per_page=per_page, headers=_headers, params=_params, ) @@ -246074,140 +256468,20 @@ def get_evaluation_test_case( return cast(JSON, deserialized) # type: ignore - @overload - def update_evaluation_test_case( - self, - test_case_uuid: str, - body: Optional[JSON] = None, - *, - content_type: str = "application/json", - **kwargs: Any, - ) -> JSON: - # pylint: disable=line-too-long - """Update an Evaluation Test Case. - - To update an evaluation test-case send a PUT request to - ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. - - :param test_case_uuid: Test-case UUID to update. Required. - :type test_case_uuid: str - :param body: Default value is None. - :type body: JSON - :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": { - "metric_uuids": [ - "str" # Optional. - ] - }, - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "test_case_uuid": "str" # Optional. Test-case UUID to update. - } - - # response body for status code(s): 200 - response == { - "test_case_uuid": "str", # Optional. - "version": 0 # Optional. The new verson of the test case. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @overload - def update_evaluation_test_case( - self, - test_case_uuid: str, - body: Optional[IO[bytes]] = None, - *, - content_type: str = "application/json", - **kwargs: Any, - ) -> JSON: - # pylint: disable=line-too-long - """Update an Evaluation Test Case. - - To update an evaluation test-case send a PUT request to - ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. - - :param test_case_uuid: Test-case UUID to update. Required. - :type test_case_uuid: str - :param body: Default value is None. - :type body: IO[bytes] - :keyword content_type: Body Parameter content-type. Content type parameter for binary body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "test_case_uuid": "str", # Optional. - "version": 0 # Optional. The new verson of the test case. - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - @distributed_trace - def update_evaluation_test_case( - self, - test_case_uuid: str, - body: Optional[Union[JSON, IO[bytes]]] = None, - **kwargs: Any, + def get_evaluation_run_prompt_results( + self, evaluation_run_uuid: str, prompt_id: int, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Update an Evaluation Test Case. + """Retrieve Results of an Evaluation Run Prompt. - To update an evaluation test-case send a PUT request to - ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. + To retrieve results of an evaluation run, send a GET request to + ``/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results/{prompt_id}``. - :param test_case_uuid: Test-case UUID to update. Required. - :type test_case_uuid: str - :param body: Is either a JSON type or a IO[bytes] type. Default value is None. - :type body: JSON or IO[bytes] + :param evaluation_run_uuid: Evaluation run UUID. Required. + :type evaluation_run_uuid: str + :param prompt_id: Prompt ID to get results for. Required. + :type prompt_id: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -246215,33 +256489,269 @@ def update_evaluation_test_case( Example: .. code-block:: python - # JSON input template you can fill out and use as your body input. - body = { - "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case - is executed. - "description": "str", # Optional. Description of the test case. - "metrics": { - "metric_uuids": [ - "str" # Optional. - ] - }, - "name": "str", # Optional. Name of the test case. - "star_metric": { - "metric_uuid": "str", # Optional. - "name": "str", # Optional. - "success_threshold": 0.0, # Optional. The success threshold for the - star metric. This is a value that the metric must reach to be considered - successful. - "success_threshold_pct": 0 # Optional. The success threshold for the - star metric. This is a percentage value between 0 and 100. - }, - "test_case_uuid": "str" # Optional. Test-case UUID to update. - } - # response body for status code(s): 200 response == { - "test_case_uuid": "str", # Optional. - "version": 0 # Optional. The new verson of the test case. + "prompt": { + "evaluation_trace_spans": [ + { + "created_at": "2020-02-20 00:00:00", # Optional. + When the span was created. + "input": {}, # Optional. Input data for the span + (flexible structure - can be messages array, string, etc.). + "name": "str", # Optional. Name/identifier for the + span. + "output": {}, # Optional. Output data from the span + (flexible structure - can be message, string, etc.). + "retriever_chunks": [ + { + "chunk_usage_pct": 0.0, # Optional. + The usage percentage of the chunk. + "chunk_used": bool, # Optional. + Indicates if the chunk was used in the prompt. + "index_uuid": "str", # Optional. The + index uuid (Knowledge Base) of the chunk. + "source_name": "str", # Optional. + The source name for the chunk, e.g., the file name or + document title. + "text": "str" # Optional. Text + content of the chunk. + } + ], + "span_level_metric_results": [ + { + "error_description": "str", # + Optional. Error description if the metric could not be + calculated. + "metric_name": "str", # Optional. + Metric name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value + is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The + value of the metric as a number. + "reasoning": "str", # Optional. + Reasoning of the metric result. + "string_value": "str" # Optional. + The value of the metric as a string. + } + ], + "spans": [ + { + "agent": { + "agent_type": + "AGENT_TYPE_UNSPECIFIED", # Optional. Default value is + "AGENT_TYPE_UNSPECIFIED". Agent span. Known values are: + "AGENT_TYPE_UNSPECIFIED", "AGENT_TYPE_DEFAULT", + "AGENT_TYPE_PLANNER", "AGENT_TYPE_REACT", + "AGENT_TYPE_REFLECTION", "AGENT_TYPE_ROUTER", + "AGENT_TYPE_CLASSIFIER", "AGENT_TYPE_SUPERVISOR", and + "AGENT_TYPE_JUDGE". + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "redacted_input": "str", # + Optional. Child spans - must contain between 1 and 999 + spans Allowed types: agent, llm, tool, retriever (not + workflow). + "redacted_output": "str", # + Optional. Child spans - must contain between 1 and 999 + spans Allowed types: agent, llm, tool, retriever (not + workflow). + "spans": [ + ... + ] + }, + "created_at": "2020-02-20 00:00:00", + # Optional. When the span was created. + "input": {}, # Optional. Input data + for the span (flexible structure - can be messages array, + string, etc.). + "llm": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "model": "str", # Optional. + LLM span. + "num_input_tokens": 0, # + Optional. LLM span. + "num_output_tokens": 0, # + Optional. LLM span. + "temperature": 0.0, # + Optional. LLM span. + "time_to_first_token_ns": + "str", # Optional. LLM span. + "tools": [ + {} # Optional. Tool + definitions passed to the model. + ], + "total_tokens": 0 # + Optional. LLM span. + }, + "name": "str", # Optional. + Name/identifier for the span. + "output": {}, # Optional. Output + data from the span (flexible structure - can be message, + string, etc.). + "retriever": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + } + }, + "tool": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "tool_call_id": "str" # + Optional. Tool span. + }, + "type": "TRACE_SPAN_TYPE_UNKNOWN", # + Optional. Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types + of spans in a trace. Known values are: + "TRACE_SPAN_TYPE_UNKNOWN", "TRACE_SPAN_TYPE_LLM", + "TRACE_SPAN_TYPE_RETRIEVER", "TRACE_SPAN_TYPE_TOOL", + "TRACE_SPAN_TYPE_AGENT", and "TRACE_SPAN_TYPE_WORKFLOW". + "workflow": { + "common": { + "created_at": + "2020-02-20 00:00:00", # Optional. Common optional + fields shared by all span types. + "duration_ns": "str", + # Optional. Common optional fields shared by all span + types. + "metadata": { + "str": "str" + # Optional. Arbitrary structured metadata. + }, + "status_code": 0, # + Optional. Common optional fields shared by all span + types. + "tags": [ + "str" # + Optional. Free-form tags for filtering/grouping. + ] + }, + "spans": [ + ... + ] + } + } + ], + "type": "TRACE_SPAN_TYPE_UNKNOWN" # Optional. + Default value is "TRACE_SPAN_TYPE_UNKNOWN". Types of spans in a + trace. Known values are: "TRACE_SPAN_TYPE_UNKNOWN", + "TRACE_SPAN_TYPE_LLM", "TRACE_SPAN_TYPE_RETRIEVER", + "TRACE_SPAN_TYPE_TOOL", "TRACE_SPAN_TYPE_AGENT", and + "TRACE_SPAN_TYPE_WORKFLOW". + } + ], + "ground_truth": "str", # Optional. The ground truth for the prompt. + "input": "str", # Optional. + "input_tokens": "str", # Optional. The number of input tokens used + in the prompt. + "output": "str", # Optional. + "output_tokens": "str", # Optional. The number of output tokens used + in the prompt. + "prompt_chunks": [ + { + "chunk_usage_pct": 0.0, # Optional. The usage + percentage of the chunk. + "chunk_used": bool, # Optional. Indicates if the + chunk was used in the prompt. + "index_uuid": "str", # Optional. The index uuid + (Knowledge Base) of the chunk. + "source_name": "str", # Optional. The source name + for the chunk, e.g., the file name or document title. + "text": "str" # Optional. Text content of the chunk. + } + ], + "prompt_id": 0, # Optional. Prompt ID. + "prompt_level_metric_results": [ + { + "error_description": "str", # Optional. Error + description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the + metric as a number. + "reasoning": "str", # Optional. Reasoning of the + metric result. + "string_value": "str" # Optional. The value of the + metric as a string. + } + ], + "trace_id": "str" # Optional. The trace id for the prompt. + } } # response body for status code(s): 404 response == { @@ -246268,30 +256778,14 @@ def update_evaluation_test_case( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} - content_type: Optional[str] = kwargs.pop( - "content_type", _headers.pop("Content-Type", None) - ) cls: ClsType[JSON] = kwargs.pop("cls", None) - content_type = content_type or "application/json" - _json = None - _content = None - if isinstance(body, (IOBase, bytes)): - _content = body - else: - if body is not None: - _json = body - else: - _json = None - - _request = build_genai_update_evaluation_test_case_request( - test_case_uuid=test_case_uuid, - content_type=content_type, - json=_json, - content=_content, + _request = build_genai_get_evaluation_run_prompt_results_request( + evaluation_run_uuid=evaluation_run_uuid, + prompt_id=prompt_id, headers=_headers, params=_params, ) @@ -246351,23 +256845,12 @@ def update_evaluation_test_case( return cast(JSON, deserialized) # type: ignore @distributed_trace - def list_indexing_jobs( - self, - *, - page: Optional[int] = None, - per_page: Optional[int] = None, - **kwargs: Any, - ) -> JSON: + def list_evaluation_test_cases(self, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Indexing Jobs for a Knowledge Base. + """List Evaluation Test Cases. - To list all indexing jobs for a knowledge base, send a GET request to - ``/v2/gen-ai/indexing_jobs``. + To list all evaluation test cases, send a GET request to ``/v2/gen-ai/evaluation_test_cases``. - :keyword page: Page number. Default value is None. - :paramtype page: int - :keyword per_page: Items per page. Default value is None. - :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -246377,97 +256860,136 @@ def list_indexing_jobs( # response body for status code(s): 200 response == { - "jobs": [ + "evaluation_test_cases": [ { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ + "archived_at": "2020-02-20 00:00:00", # Optional. + Alternative way of authentication for internal usage only - should not be + exposed to public api. + "created_at": "2020-02-20 00:00:00", # Optional. Alternative + way of authentication for internal usage only - should not be exposed to + public api. + "created_by_user_email": "str", # Optional. Alternative way + of authentication for internal usage only - should not be exposed to + public api. + "created_by_user_id": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "dataset": { + "created_at": "2020-02-20 00:00:00", # Optional. + Time created at. + "dataset_name": "str", # Optional. Name of the + dataset. + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # + Optional. Default value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known + values are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". + "dataset_uuid": "str", # Optional. UUID of the + dataset. + "file_size": "str", # Optional. The size of the + dataset uploaded file in bytes. + "has_ground_truth": bool, # Optional. Does the + dataset have a ground truth column?. + "row_count": 0 # Optional. Number of rows in the + dataset. + }, + "dataset_name": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "dataset_uuid": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "description": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "latest_version_number_of_runs": 0, # Optional. Alternative + way of authentication for internal usage only - should not be exposed to + public api. + "metrics": [ { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. + "category": "METRIC_CATEGORY_UNSPECIFIED", # + Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known + values are: "METRIC_CATEGORY_UNSPECIFIED", + "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", + "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and + "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether a + metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the + metric is inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "metric_name": "str", # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "metric_rank": 0, # Optional. Alternative + way of authentication for internal usage only - should not be + exposed to public api. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known + values are: "METRIC_TYPE_UNSPECIFIED", + "METRIC_TYPE_GENERAL_QUALITY", "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. + Alternative way of authentication for internal usage only - + should not be exposed to public api. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum + value for the metric. + "range_min": 0.0 # Optional. The minimum + value for the metric. } ], - "data_source_uuids": [ - "str" # Optional. The indexing jobs. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - } - ], - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "name": "str", # Optional. Alternative way of authentication + for internal usage only - should not be exposed to public api. + "star_metric": { + "metric_uuid": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to + public api. + "name": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to + public api. + "success_threshold": 0.0, # Optional. The success + threshold for the star metric. This is a value that the metric must + reach to be considered successful. + "success_threshold_pct": 0 # Optional. The success + threshold for the star metric. This is a percentage value between 0 + and 100. + }, + "test_case_uuid": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "total_runs": 0, # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "updated_at": "2020-02-20 00:00:00", # Optional. Alternative + way of authentication for internal usage only - should not be exposed to + public api. + "updated_by_user_email": "str", # Optional. Alternative way + of authentication for internal usage only - should not be exposed to + public api. + "updated_by_user_id": "str", # Optional. Alternative way of + authentication for internal usage only - should not be exposed to public + api. + "version": 0 # Optional. Alternative way of authentication + for internal usage only - should not be exposed to public api. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. - "total": 0 # Optional. Total amount of items over all pages. - } + ] } # response body for status code(s): 404 response == { @@ -246499,9 +257021,7 @@ def list_indexing_jobs( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_indexing_jobs_request( - page=page, - per_page=per_page, + _request = build_genai_list_evaluation_test_cases_request( headers=_headers, params=_params, ) @@ -246561,7 +257081,7 @@ def list_indexing_jobs( return cast(JSON, deserialized) # type: ignore @overload - def create_indexing_job( + def create_evaluation_test_case( self, body: Optional[JSON] = None, *, @@ -246569,10 +257089,9 @@ def create_indexing_job( **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Start Indexing Job for a Knowledge Base. + """Create Evaluation Test Case. - To start an indexing job for a knowledge base, send a POST request to - ``/v2/gen-ai/indexing_jobs``. + To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. :param body: Default value is None. :type body: JSON @@ -246588,86 +257107,29 @@ def create_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "data_source_uuids": [ - "str" # Optional. List of data source ids to index, if none are - provided, all data sources will be indexed. + "agent_workspace_name": "str", # Optional. + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": [ + "str" # Optional. Full metric list to use for evaluation test case. ], - "knowledge_base_uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "workspace_uuid": "str" # Optional. The workspace uuid. } # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. - } + "test_case_uuid": "str" # Optional. Test"u2011case UUID. } # response body for status code(s): 404 response == { @@ -246683,7 +257145,7 @@ def create_indexing_job( """ @overload - def create_indexing_job( + def create_evaluation_test_case( self, body: Optional[IO[bytes]] = None, *, @@ -246691,10 +257153,9 @@ def create_indexing_job( **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Start Indexing Job for a Knowledge Base. + """Create Evaluation Test Case. - To start an indexing job for a knowledge base, send a POST request to - ``/v2/gen-ai/indexing_jobs``. + To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. :param body: Default value is None. :type body: IO[bytes] @@ -246710,77 +257171,7 @@ def create_indexing_job( # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. - } + "test_case_uuid": "str" # Optional. Test"u2011case UUID. } # response body for status code(s): 404 response == { @@ -246796,14 +257187,13 @@ def create_indexing_job( """ @distributed_trace - def create_indexing_job( + def create_evaluation_test_case( self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Start Indexing Job for a Knowledge Base. + """Create Evaluation Test Case. - To start an indexing job for a knowledge base, send a POST request to - ``/v2/gen-ai/indexing_jobs``. + To create an evaluation test-case send a POST request to ``/v2/gen-ai/evaluation_test_cases``. :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] @@ -246816,86 +257206,29 @@ def create_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "data_source_uuids": [ - "str" # Optional. List of data source ids to index, if none are - provided, all data sources will be indexed. + "agent_workspace_name": "str", # Optional. + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": [ + "str" # Optional. Full metric list to use for evaluation test case. ], - "knowledge_base_uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "workspace_uuid": "str" # Optional. The workspace uuid. } # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. - } + "test_case_uuid": "str" # Optional. Test"u2011case UUID. } # response body for status code(s): 404 response == { @@ -246941,7 +257274,7 @@ def create_indexing_job( else: _json = None - _request = build_genai_create_indexing_job_request( + _request = build_genai_create_evaluation_test_case_request( content_type=content_type, json=_json, content=_content, @@ -247004,17 +257337,23 @@ def create_indexing_job( return cast(JSON, deserialized) # type: ignore @distributed_trace - def list_indexing_job_data_sources( - self, indexing_job_uuid: str, **kwargs: Any + def list_evaluation_runs_by_test_case( + self, + evaluation_test_case_uuid: str, + *, + evaluation_test_case_version: Optional[int] = None, + **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """List Data Sources for Indexing Job for a Knowledge Base. + """List Evaluation Runs by Test Case. - To list all datasources for an indexing job, send a GET request to - ``/v2/gen-ai/indexing_jobs/{indexing_job_uuid}/data_sources``. + To list all evaluation runs by test case, send a GET request to + ``/v2/gen-ai/evaluation_test_cases/{evaluation_test_case_uuid}/evaluation_runs``. - :param indexing_job_uuid: Uuid of the indexing job. Required. - :type indexing_job_uuid: str + :param evaluation_test_case_uuid: Evaluation run UUID. Required. + :type evaluation_test_case_uuid: str + :keyword evaluation_test_case_version: Version of the test case. Default value is None. + :paramtype evaluation_test_case_version: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -247024,40 +257363,82 @@ def list_indexing_job_data_sources( # response body for status code(s): 200 response == { - "indexed_data_sources": [ + "evaluation_runs": [ { - "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp - when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the indexed - data source. - "error_details": "str", # Optional. A detailed error + "agent_deleted": bool, # Optional. Whether agent is deleted. + "agent_deployment_name": "str", # Optional. The agent + deployment name. + "agent_name": "str", # Optional. Agent name. + "agent_uuid": "str", # Optional. Agent UUID. + "agent_version_hash": "str", # Optional. Version hash. + "agent_workspace_uuid": "str", # Optional. Agent workspace + uuid. + "created_by_user_email": "str", # Optional. List of + evaluation runs. + "created_by_user_id": "str", # Optional. List of evaluation + runs. + "error_description": "str", # Optional. The error description. - "error_msg": "str", # Optional. A string code provinding a - hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count of files - that have failed. - "indexed_file_count": "str", # Optional. Total count of - files that have been indexed. - "indexed_item_count": "str", # Optional. Total count of - files that have been indexed. - "removed_item_count": "str", # Optional. Total count of - files that have been removed. - "skipped_item_count": "str", # Optional. Total count of - files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. Timestamp - when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default - value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of files in - data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size of - files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file count in - the data source. + "evaluation_run_uuid": "str", # Optional. Evaluation run + UUID. + "evaluation_test_case_workspace_uuid": "str", # Optional. + Evaluation test case workspace uuid. + "finished_at": "2020-02-20 00:00:00", # Optional. Run end + time. + "pass_status": bool, # Optional. The pass status of the + evaluation run based on the star metric. + "queued_at": "2020-02-20 00:00:00", # Optional. Run queued + time. + "run_level_metric_results": [ + { + "error_description": "str", # Optional. + Error description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric + name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value + of the metric as a number. + "reasoning": "str", # Optional. Reasoning of + the metric result. + "string_value": "str" # Optional. The value + of the metric as a string. + } + ], + "run_name": "str", # Optional. Run name. + "star_metric_result": { + "error_description": "str", # Optional. Error + description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric name. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value of the + metric as a number. + "reasoning": "str", # Optional. Reasoning of the + metric result. + "string_value": "str" # Optional. The value of the + metric as a string. + }, + "started_at": "2020-02-20 00:00:00", # Optional. Run start + time. + "status": "EVALUATION_RUN_STATUS_UNSPECIFIED", # Optional. + Default value is "EVALUATION_RUN_STATUS_UNSPECIFIED". Evaluation Run + Statuses. Known values are: "EVALUATION_RUN_STATUS_UNSPECIFIED", + "EVALUATION_RUN_QUEUED", "EVALUATION_RUN_RUNNING_DATASET", + "EVALUATION_RUN_EVALUATING_RESULTS", "EVALUATION_RUN_CANCELLING", + "EVALUATION_RUN_CANCELLED", "EVALUATION_RUN_SUCCESSFUL", + "EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and "EVALUATION_RUN_FAILED". + "test_case_description": "str", # Optional. Test case + description. + "test_case_name": "str", # Optional. Test case name. + "test_case_uuid": "str", # Optional. Test-case UUID. + "test_case_version": 0 # Optional. Test-case-version. } ] } @@ -247091,8 +257472,9 @@ def list_indexing_job_data_sources( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_indexing_job_data_sources_request( - indexing_job_uuid=indexing_job_uuid, + _request = build_genai_list_evaluation_runs_by_test_case_request( + evaluation_test_case_uuid=evaluation_test_case_uuid, + evaluation_test_case_version=evaluation_test_case_version, headers=_headers, params=_params, ) @@ -247152,17 +257534,23 @@ def list_indexing_job_data_sources( return cast(JSON, deserialized) # type: ignore @distributed_trace - def get_indexing_job_details_signed_url( - self, indexing_job_uuid: str, **kwargs: Any + def get_evaluation_test_case( + self, + test_case_uuid: str, + *, + evaluation_test_case_version: Optional[int] = None, + **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Get Signed URL for Indexing Job Details. + """Retrieve Information About an Existing Evaluation Test Case. - To get a signed URL for indexing job details, send a GET request to - ``/v2/gen-ai/indexing_jobs/{uuid}/details_signed_url``. + To retrive information about an existing evaluation test case, send a GET request to + ``/v2/gen-ai/evaluation_test_case/{test_case_uuid}``. - :param indexing_job_uuid: The uuid of the indexing job. Required. - :type indexing_job_uuid: str + :param test_case_uuid: The test case uuid to retrieve. Required. + :type test_case_uuid: str + :keyword evaluation_test_case_version: Version of the test case. Default value is None. + :paramtype evaluation_test_case_version: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -247172,8 +257560,86 @@ def get_indexing_job_details_signed_url( # response body for status code(s): 200 response == { - "signed_url": "str" # Optional. The signed url for downloading the indexing - job details. + "evaluation_test_case": { + "archived_at": "2020-02-20 00:00:00", # Optional. + "created_at": "2020-02-20 00:00:00", # Optional. + "created_by_user_email": "str", # Optional. + "created_by_user_id": "str", # Optional. + "dataset": { + "created_at": "2020-02-20 00:00:00", # Optional. Time + created at. + "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # + Optional. Default value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known + values are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". + "dataset_uuid": "str", # Optional. UUID of the dataset. + "file_size": "str", # Optional. The size of the dataset + uploaded file in bytes. + "has_ground_truth": bool, # Optional. Does the dataset have + a ground truth column?. + "row_count": 0 # Optional. Number of rows in the dataset. + }, + "dataset_name": "str", # Optional. + "dataset_uuid": "str", # Optional. + "description": "str", # Optional. + "latest_version_number_of_runs": 0, # Optional. + "metrics": [ + { + "category": "METRIC_CATEGORY_UNSPECIFIED", # + Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known + values are: "METRIC_CATEGORY_UNSPECIFIED", + "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", + "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", + # Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope + that determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to + agent metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. + "metric_name": "str", # Optional. + "metric_rank": 0, # Optional. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values + are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for + the metric. + "range_min": 0.0 # Optional. The minimum value for + the metric. + } + ], + "name": "str", # Optional. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold + for the star metric. This is a value that the metric must reach to be + considered successful. + "success_threshold_pct": 0 # Optional. The success threshold + for the star metric. This is a percentage value between 0 and 100. + }, + "test_case_uuid": "str", # Optional. + "total_runs": 0, # Optional. + "updated_at": "2020-02-20 00:00:00", # Optional. + "updated_by_user_email": "str", # Optional. + "updated_by_user_id": "str", # Optional. + "version": 0 # Optional. + } } # response body for status code(s): 404 response == { @@ -247205,8 +257671,9 @@ def get_indexing_job_details_signed_url( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_indexing_job_details_signed_url_request( - indexing_job_uuid=indexing_job_uuid, + _request = build_genai_get_evaluation_test_case_request( + test_case_uuid=test_case_uuid, + evaluation_test_case_version=evaluation_test_case_version, headers=_headers, params=_params, ) @@ -247265,16 +257732,140 @@ def get_indexing_job_details_signed_url( return cast(JSON, deserialized) # type: ignore + @overload + def update_evaluation_test_case( + self, + test_case_uuid: str, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Update an Evaluation Test Case. + + To update an evaluation test-case send a PUT request to + ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. + + :param test_case_uuid: Test-case UUID to update. Required. + :type test_case_uuid: str + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": { + "metric_uuids": [ + "str" # Optional. + ] + }, + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "test_case_uuid": "str" # Optional. Test-case UUID to update. + } + + # response body for status code(s): 200 + response == { + "test_case_uuid": "str", # Optional. + "version": 0 # Optional. The new verson of the test case. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + def update_evaluation_test_case( + self, + test_case_uuid: str, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Update an Evaluation Test Case. + + To update an evaluation test-case send a PUT request to + ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. + + :param test_case_uuid: Test-case UUID to update. Required. + :type test_case_uuid: str + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "test_case_uuid": "str", # Optional. + "version": 0 # Optional. The new verson of the test case. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + @distributed_trace - def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: + def update_evaluation_test_case( + self, + test_case_uuid: str, + body: Optional[Union[JSON, IO[bytes]]] = None, + **kwargs: Any, + ) -> JSON: # pylint: disable=line-too-long - """Retrieve Status of Indexing Job for a Knowledge Base. + """Update an Evaluation Test Case. - To get status of an indexing Job for a knowledge base, send a GET request to - ``/v2/gen-ai/indexing_jobs/{uuid}``. + To update an evaluation test-case send a PUT request to + ``/v2/gen-ai/evaluation_test_cases/{test_case_uuid}``. - :param uuid: Indexing job id. Required. - :type uuid: str + :param test_case_uuid: Test-case UUID to update. Required. + :type test_case_uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -247282,78 +257873,258 @@ def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "dataset_uuid": "str", # Optional. Dataset against which the test"u2011case + is executed. + "description": "str", # Optional. Description of the test case. + "metrics": { + "metric_uuids": [ + "str" # Optional. + ] + }, + "name": "str", # Optional. Name of the test case. + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + }, + "test_case_uuid": "str" # Optional. Test-case UUID to update. + } + # response body for status code(s): 200 response == { - "job": { - "completed_datasources": 0, # Optional. Number of datasources - indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid of the - indexed data source. - "error_details": "str", # Optional. A detailed error - description. - "error_msg": "str", # Optional. A string code - provinding a hint which part of the system experienced an error. - "failed_item_count": "str", # Optional. Total count - of files that have failed. - "indexed_file_count": "str", # Optional. Total count - of files that have been indexed. - "indexed_item_count": "str", # Optional. Total count - of files that have been indexed. - "removed_item_count": "str", # Optional. Total count - of files that have been removed. - "skipped_item_count": "str", # Optional. Total count - of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # Optional. - Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. - Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size of - files in data source in bytes. - "total_bytes_indexed": "str", # Optional. Total size - of files in data source in bytes that have been indexed. - "total_file_count": "str" # Optional. Total file - count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "is_report_available": bool, # Optional. Boolean value to determine - if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is - "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", - "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", - "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", - "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is - "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", - "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", - "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", - "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and - "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources being - indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By the - Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "uuid": "str" # Optional. Unique id. + "test_case_uuid": "str", # Optional. + "version": 0 # Optional. The new verson of the test case. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + cls: ClsType[JSON] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_update_evaluation_test_case_request( + test_case_uuid=test_case_uuid, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def list_indexing_jobs( + self, + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """List Indexing Jobs for a Knowledge Base. + + To list all indexing jobs for a knowledge base, send a GET request to + ``/v2/gen-ai/indexing_jobs``. + + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "jobs": [ + { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. The indexing jobs. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + } + ], + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. } } # response body for status code(s): 404 @@ -247386,8 +258157,9 @@ def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_indexing_job_request( - uuid=uuid, + _request = build_genai_list_indexing_jobs_request( + page=page, + per_page=per_page, headers=_headers, params=_params, ) @@ -247447,22 +258219,19 @@ def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - def cancel_indexing_job( + def create_indexing_job( self, - uuid: str, body: Optional[JSON] = None, *, content_type: str = "application/json", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Cancel Indexing Job for a Knowledge Base. + """Start Indexing Job for a Knowledge Base. - To cancel an indexing job for a knowledge base, send a PUT request to - ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + To start an indexing job for a knowledge base, send a POST request to + ``/v2/gen-ai/indexing_jobs``. - :param uuid: A unique identifier for an indexing job. Required. - :type uuid: str :param body: Default value is None. :type body: JSON :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. @@ -247477,7 +258246,11 @@ def cancel_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "uuid": "str" # Optional. A unique identifier for an indexing job. + "data_source_uuids": [ + "str" # Optional. List of data source ids to index, if none are + provided, all data sources will be indexed. + ], + "knowledge_base_uuid": "str" # Optional. Knowledge base id. } # response body for status code(s): 200 @@ -247568,22 +258341,19 @@ def cancel_indexing_job( """ @overload - def cancel_indexing_job( + def create_indexing_job( self, - uuid: str, body: Optional[IO[bytes]] = None, *, content_type: str = "application/json", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Cancel Indexing Job for a Knowledge Base. + """Start Indexing Job for a Knowledge Base. - To cancel an indexing job for a knowledge base, send a PUT request to - ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + To start an indexing job for a knowledge base, send a POST request to + ``/v2/gen-ai/indexing_jobs``. - :param uuid: A unique identifier for an indexing job. Required. - :type uuid: str :param body: Default value is None. :type body: IO[bytes] :keyword content_type: Body Parameter content-type. Content type parameter for binary body. @@ -247684,17 +258454,15 @@ def cancel_indexing_job( """ @distributed_trace - def cancel_indexing_job( - self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + def create_indexing_job( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Cancel Indexing Job for a Knowledge Base. + """Start Indexing Job for a Knowledge Base. - To cancel an indexing job for a knowledge base, send a PUT request to - ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + To start an indexing job for a knowledge base, send a POST request to + ``/v2/gen-ai/indexing_jobs``. - :param uuid: A unique identifier for an indexing job. Required. - :type uuid: str :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] :return: JSON object @@ -247706,7 +258474,11 @@ def cancel_indexing_job( # JSON input template you can fill out and use as your body input. body = { - "uuid": "str" # Optional. A unique identifier for an indexing job. + "data_source_uuids": [ + "str" # Optional. List of data source ids to index, if none are + provided, all data sources will be indexed. + ], + "knowledge_base_uuid": "str" # Optional. Knowledge base id. } # response body for status code(s): 200 @@ -247827,8 +258599,7 @@ def cancel_indexing_job( else: _json = None - _request = build_genai_cancel_indexing_job_request( - uuid=uuid, + _request = build_genai_create_indexing_job_request( content_type=content_type, json=_json, content=_content, @@ -247891,22 +258662,17 @@ def cancel_indexing_job( return cast(JSON, deserialized) # type: ignore @distributed_trace - def list_knowledge_bases( - self, - *, - page: Optional[int] = None, - per_page: Optional[int] = None, - **kwargs: Any, + def list_indexing_job_data_sources( + self, indexing_job_uuid: str, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """List Knowledge Bases. + """List Data Sources for Indexing Job for a Knowledge Base. - To list all knowledge bases, send a GET request to ``/v2/gen-ai/knowledge_bases``. + To list all datasources for an indexing job, send a GET request to + ``/v2/gen-ai/indexing_jobs/{indexing_job_uuid}/data_sources``. - :keyword page: Page number. Default value is None. - :paramtype page: int - :keyword per_page: Items per page. Default value is None. - :paramtype per_page: int + :param indexing_job_uuid: Uuid of the indexing job. Required. + :type indexing_job_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -247916,129 +258682,1021 @@ def list_knowledge_bases( # response body for status code(s): 200 response == { - "knowledge_bases": [ + "indexed_data_sources": [ { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time - when the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "database_id": "str", # Optional. The knowledge bases. - "embedding_model_uuid": "str", # Optional. The knowledge - bases. - "is_public": bool, # Optional. Whether the knowledge base is - public or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. - Creation date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 - 00:00:00", # Optional. Timestamp when data source completed - indexing. - "data_source_uuid": "str", # - Optional. Uuid of the indexed data source. - "error_details": "str", # Optional. - A detailed error description. - "error_msg": "str", # Optional. A - string code provinding a hint which part of the system - experienced an error. - "failed_item_count": "str", # - Optional. Total count of files that have failed. - "indexed_file_count": "str", # - Optional. Total count of files that have been indexed. - "indexed_item_count": "str", # - Optional. Total count of files that have been indexed. - "removed_item_count": "str", # - Optional. Total count of files that have been removed. - "skipped_item_count": "str", # - Optional. Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", - # Optional. Timestamp when data source started indexing. - "status": - "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default value is - "DATA_SOURCE_STATUS_UNKNOWN". Known values are: - "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", - "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", - "DATA_SOURCE_STATUS_FAILED", and - "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. - Total size of files in data source in bytes. - "total_bytes_indexed": "str", # - Optional. Total size of files in data source in bytes that - have been indexed. - "total_file_count": "str" # - Optional. Total file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean - value to determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge - base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. - Default value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. - Default value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This - field is deprecated]. - "total_datasources": 0, # Optional. Number of - datasources being indexed. - "total_tokens": "str", # Optional. Total Tokens - Consumed By the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. - Last modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. The knowledge bases. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is - enabled for retrieval. - "model": "str" # Optional. Reranker model internal - name. - }, - "tags": [ - "str" # Optional. Tags to organize related - resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "user_id": "str", # Optional. Id of user that created the - knowledge base. - "uuid": "str" # Optional. Unique id for knowledge base. - } - ], - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "completed_at": "2020-02-20 00:00:00", # Optional. Timestamp + when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the indexed + data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code provinding a + hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count of files + that have failed. + "indexed_file_count": "str", # Optional. Total count of + files that have been indexed. + "indexed_item_count": "str", # Optional. Total count of + files that have been indexed. + "removed_item_count": "str", # Optional. Total count of + files that have been removed. + "skipped_item_count": "str", # Optional. Total count of + files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. Timestamp + when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default + value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of files in + data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size of + files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file count in + the data source. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_list_indexing_job_data_sources_request( + indexing_job_uuid=indexing_job_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def get_indexing_job_details_signed_url( + self, indexing_job_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Get Signed URL for Indexing Job Details. + + To get a signed URL for indexing job details, send a GET request to + ``/v2/gen-ai/indexing_jobs/{uuid}/details_signed_url``. + + :param indexing_job_uuid: The uuid of the indexing job. Required. + :type indexing_job_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "signed_url": "str" # Optional. The signed url for downloading the indexing + job details. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_indexing_job_details_signed_url_request( + indexing_job_uuid=indexing_job_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def get_indexing_job(self, uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Retrieve Status of Indexing Job for a Knowledge Base. + + To get status of an indexing Job for a knowledge base, send a GET request to + ``/v2/gen-ai/indexing_jobs/{uuid}``. + + :param uuid: Indexing job id. Required. + :type uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_indexing_job_request( + uuid=uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @overload + def cancel_indexing_job( + self, + uuid: str, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Cancel Indexing Job for a Knowledge Base. + + To cancel an indexing job for a knowledge base, send a PUT request to + ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + + :param uuid: A unique identifier for an indexing job. Required. + :type uuid: str + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "uuid": "str" # Optional. A unique identifier for an indexing job. + } + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + def cancel_indexing_job( + self, + uuid: str, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Cancel Indexing Job for a Knowledge Base. + + To cancel an indexing job for a knowledge base, send a PUT request to + ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + + :param uuid: A unique identifier for an indexing job. Required. + :type uuid: str + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace + def cancel_indexing_job( + self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Cancel Indexing Job for a Knowledge Base. + + To cancel an indexing job for a knowledge base, send a PUT request to + ``/v2/gen-ai/indexing_jobs/{uuid}/cancel``. + + :param uuid: A unique identifier for an indexing job. Required. + :type uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "uuid": "str" # Optional. A unique identifier for an indexing job. + } + + # response body for status code(s): 200 + response == { + "job": { + "completed_datasources": 0, # Optional. Number of datasources + indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid of the + indexed data source. + "error_details": "str", # Optional. A detailed error + description. + "error_msg": "str", # Optional. A string code + provinding a hint which part of the system experienced an error. + "failed_item_count": "str", # Optional. Total count + of files that have failed. + "indexed_file_count": "str", # Optional. Total count + of files that have been indexed. + "indexed_item_count": "str", # Optional. Total count + of files that have been indexed. + "removed_item_count": "str", # Optional. Total count + of files that have been removed. + "skipped_item_count": "str", # Optional. Total count + of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # Optional. + Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # Optional. + Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size of + files in data source in bytes. + "total_bytes_indexed": "str", # Optional. Total size + of files in data source in bytes that have been indexed. + "total_file_count": "str" # Optional. Total file + count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "is_report_available": bool, # Optional. Boolean value to determine + if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default value is + "BATCH_JOB_PHASE_UNKNOWN". Known values are: "BATCH_JOB_PHASE_UNKNOWN", + "BATCH_JOB_PHASE_PENDING", "BATCH_JOB_PHASE_RUNNING", + "BATCH_JOB_PHASE_SUCCEEDED", "BATCH_JOB_PHASE_FAILED", + "BATCH_JOB_PHASE_ERROR", and "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default value is + "INDEX_JOB_STATUS_UNKNOWN". Known values are: "INDEX_JOB_STATUS_UNKNOWN", + "INDEX_JOB_STATUS_PARTIAL", "INDEX_JOB_STATUS_IN_PROGRESS", + "INDEX_JOB_STATUS_COMPLETED", "INDEX_JOB_STATUS_FAILED", + "INDEX_JOB_STATUS_NO_CHANGES", "INDEX_JOB_STATUS_PENDING", and + "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources being + indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By the + Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + cls: ClsType[JSON] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_cancel_indexing_job_request( + uuid=uuid, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def list_knowledge_bases( + self, + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """List Knowledge Bases. + + To list all knowledge bases, send a GET request to ``/v2/gen-ai/knowledge_bases``. + + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "knowledge_bases": [ + { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time + when the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "database_id": "str", # Optional. The knowledge bases. + "embedding_model_uuid": "str", # Optional. The knowledge + bases. + "is_public": bool, # Optional. Whether the knowledge base is + public or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 + 00:00:00", # Optional. Timestamp when data source completed + indexing. + "data_source_uuid": "str", # + Optional. Uuid of the indexed data source. + "error_details": "str", # Optional. + A detailed error description. + "error_msg": "str", # Optional. A + string code provinding a hint which part of the system + experienced an error. + "failed_item_count": "str", # + Optional. Total count of files that have failed. + "indexed_file_count": "str", # + Optional. Total count of files that have been indexed. + "indexed_item_count": "str", # + Optional. Total count of files that have been indexed. + "removed_item_count": "str", # + Optional. Total count of files that have been removed. + "skipped_item_count": "str", # + Optional. Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", + # Optional. Timestamp when data source started indexing. + "status": + "DATA_SOURCE_STATUS_UNKNOWN", # Optional. Default value is + "DATA_SOURCE_STATUS_UNKNOWN". Known values are: + "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", + "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", + "DATA_SOURCE_STATUS_FAILED", and + "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. + Total size of files in data source in bytes. + "total_bytes_indexed": "str", # + Optional. Total size of files in data source in bytes that + have been indexed. + "total_file_count": "str" # + Optional. Total file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean + value to determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge + base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. + Default value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. + Default value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This + field is deprecated]. + "total_datasources": 0, # Optional. Number of + datasources being indexed. + "total_tokens": "str", # Optional. Total Tokens + Consumed By the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. The knowledge bases. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is + enabled for retrieval. + "model": "str" # Optional. Reranker model internal + name. + }, + "tags": [ + "str" # Optional. Tags to organize related + resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "user_id": "str", # Optional. Id of user that created the + knowledge base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + ], + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. "total": 0 # Optional. Total amount of items over all pages. } } @@ -248177,18 +259835,21 @@ def create_knowledge_base( data_source_details. "bucket_region": "str", # Optional. Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The - chunking algorithm to use for processing data sources. **Note: This - feature requires enabling the knowledgebase enhancements feature preview - flag.**. Known values are: "CHUNKING_ALGORITHM_UNKNOWN", - "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", - "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # + Optional. Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values + are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and + "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical - options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Optional data + sources to attach at creation. Omit or use an empty list to create + the knowledge base without sources, then add sources (with chunking + strategy and sizes) using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic @@ -248215,10 +259876,14 @@ def create_knowledge_base( you can obrain a refresh token by following the oauth2 flow. see /v2/gen-ai/oauth2/google/tokens for reference. }, - "item_path": "str", # Optional. The data sources to use for - this knowledge base. See `Organize Data Sources - `_ - for more information on data sources best practices. + "item_path": "str", # Optional. Optional data sources to + attach at creation. Omit or use an empty list to create the knowledge + base without sources, then add sources (with chunking strategy and sizes) + using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. "spaces_data_source": { "bucket_name": "str", # Optional. Spaces bucket name. @@ -248260,6 +259925,11 @@ def create_knowledge_base( retrieval. "model": "str" # Optional. Reranker model internal name. }, + "size": "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", # Optional. Default value is + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED". Known values are: + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", "OPEN_SEARCH_PLAN_SIZE_SMALL", + "OPEN_SEARCH_PLAN_SIZE_MEDIUM", "OPEN_SEARCH_PLAN_SIZE_LARGE", and + "OPEN_SEARCH_PLAN_SIZE_EXTRA_LARGE". "tags": [ "str" # Optional. Tags to organize your knowledge base. ], @@ -248565,18 +260235,21 @@ def create_knowledge_base( data_source_details. "bucket_region": "str", # Optional. Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The - chunking algorithm to use for processing data sources. **Note: This - feature requires enabling the knowledgebase enhancements feature preview - flag.**. Known values are: "CHUNKING_ALGORITHM_UNKNOWN", - "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", - "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # + Optional. Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values + are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and + "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical - options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Optional data + sources to attach at creation. Omit or use an empty list to create + the knowledge base without sources, then add sources (with chunking + strategy and sizes) using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic @@ -248603,10 +260276,14 @@ def create_knowledge_base( you can obrain a refresh token by following the oauth2 flow. see /v2/gen-ai/oauth2/google/tokens for reference. }, - "item_path": "str", # Optional. The data sources to use for - this knowledge base. See `Organize Data Sources - `_ - for more information on data sources best practices. + "item_path": "str", # Optional. Optional data sources to + attach at creation. Omit or use an empty list to create the knowledge + base without sources, then add sources (with chunking strategy and sizes) + using `Add a Data Source to a Knowledge Base + <#operation/create_knowledge_base_data_source>`_. When provided, see + `Organize Data Sources + `_ + for best practices. "spaces_data_source": { "bucket_name": "str", # Optional. Spaces bucket name. @@ -248648,6 +260325,11 @@ def create_knowledge_base( retrieval. "model": "str" # Optional. Reranker model internal name. }, + "size": "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", # Optional. Default value is + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED". Known values are: + "OPEN_SEARCH_PLAN_SIZE_UNSPECIFIED", "OPEN_SEARCH_PLAN_SIZE_SMALL", + "OPEN_SEARCH_PLAN_SIZE_MEDIUM", "OPEN_SEARCH_PLAN_SIZE_LARGE", and + "OPEN_SEARCH_PLAN_SIZE_EXTRA_LARGE". "tags": [ "str" # Optional. Tags to organize your knowledge base. ], @@ -249183,18 +260865,14 @@ def list_knowledge_base_data_sources( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The - chunking algorithm to use for processing data sources. **Note: This - feature requires enabling the knowledgebase enhancements feature preview - flag.**. Known values are: "CHUNKING_ALGORITHM_UNKNOWN", - "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", - "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # + Optional. Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values + are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and + "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical - options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. The data sources. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic @@ -249433,17 +261111,14 @@ def create_knowledge_base_data_source( "region": "str", # Optional. Region of bucket. "secret_key": "str" # Optional. The AWS Secret Key. }, - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -249483,17 +261158,15 @@ def create_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -249633,17 +261306,15 @@ def create_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -249776,17 +261447,14 @@ def create_knowledge_base_data_source( "region": "str", # Optional. Region of bucket. "secret_key": "str" # Optional. The AWS Secret Key. }, - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -249826,17 +261494,15 @@ def create_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -250064,17 +261730,14 @@ def update_knowledge_base_data_source( # JSON input template you can fill out and use as your body input. body = { - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -250093,17 +261756,15 @@ def update_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -250246,17 +261907,15 @@ def update_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -250385,17 +262044,14 @@ def update_knowledge_base_data_source( # JSON input template you can fill out and use as your body input. body = { - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # Optional. - Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking algorithm to - use for processing data sources. **Note: This feature requires enabling the - knowledgebase enhancements feature preview flag.**. Known values are: + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. Default + value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and Fixed_Length - options. + "child_chunk_size": 0, # Optional. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -250414,17 +262070,15 @@ def update_knowledge_base_data_source( }, "bucket_name": "str", # Optional. Name of storage bucket - Deprecated, moved to data_source_details. - "chunking_algorithm": "CHUNKING_ALGORITHM_SECTION_BASED", # - Optional. Default value is "CHUNKING_ALGORITHM_SECTION_BASED". The chunking - algorithm to use for processing data sources. **Note: This feature requires - enabling the knowledgebase enhancements feature preview flag.**. Known values - are: "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", + "chunking_algorithm": "CHUNKING_ALGORITHM_UNKNOWN", # Optional. + Default value is "CHUNKING_ALGORITHM_UNKNOWN". Known values are: + "CHUNKING_ALGORITHM_UNKNOWN", "CHUNKING_ALGORITHM_SECTION_BASED", "CHUNKING_ALGORITHM_HIERARCHICAL", "CHUNKING_ALGORITHM_SEMANTIC", and "CHUNKING_ALGORITHM_FIXED_LENGTH". "chunking_options": { - "child_chunk_size": 0, # Optional. Hierarchical options. - "max_chunk_size": 0, # Optional. Section_Based and - Fixed_Length options. + "child_chunk_size": 0, # Optional. Data Source configuration + for Knowledge Bases. + "max_chunk_size": 0, # Optional. Common options. "parent_chunk_size": 0, # Optional. Hierarchical options. "semantic_threshold": 0.0 # Optional. Semantic options. }, @@ -250620,19 +262274,1346 @@ def update_knowledge_base_data_source( return cast(JSON, deserialized) # type: ignore @distributed_trace - def delete_knowledge_base_data_source( - self, knowledge_base_uuid: str, data_source_uuid: str, **kwargs: Any + def delete_knowledge_base_data_source( + self, knowledge_base_uuid: str, data_source_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Delete a Data Source from a Knowledge Base. + + To delete a data source from a knowledge base, send a DELETE request to + ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/data_sources/{data_source_uuid}``. + + :param knowledge_base_uuid: Knowledge base id. Required. + :type knowledge_base_uuid: str + :param data_source_uuid: Data source id. Required. + :type data_source_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "data_source_uuid": "str", # Optional. Data source id. + "knowledge_base_uuid": "str" # Optional. Knowledge base id. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_delete_knowledge_base_data_source_request( + knowledge_base_uuid=knowledge_base_uuid, + data_source_uuid=data_source_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def list_indexing_jobs_by_knowledge_base( + self, knowledge_base_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """List Indexing Jobs for a Knowledge Base. + + To list latest 15 indexing jobs for a knowledge base, send a GET request to + ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/indexing_jobs``. + + :param knowledge_base_uuid: Knowledge base uuid in string. Required. + :type knowledge_base_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "jobs": [ + { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. The indexing jobs. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. The + indexing jobs. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + } + ], + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_list_indexing_jobs_by_knowledge_base_request( + knowledge_base_uuid=knowledge_base_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Retrieve Information About an Existing Knowledge Base. + + To retrive information about an existing knowledge base, send a GET request to + ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "database_status": "CREATING", # Optional. Default value is "CREATING". + Known values are: "CREATING", "ONLINE", "POWEROFF", "REBUILDING", "REBALANCING", + "DECOMMISSIONED", "FORKING", "MIGRATING", "RESIZING", "RESTORING", "POWERING_ON", + "UNHEALTHY", and "UPGRADING". + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_knowledge_base_request( + uuid=uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @overload + def update_knowledge_base( + self, + uuid: str, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Update a Knowledge Base. + + To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "database_id": "str", # Optional. The id of the DigitalOcean database this + knowledge base will use, optional. + "name": "str", # Optional. Knowledge base name. + "project_id": "str", # Optional. The id of the DigitalOcean project this + knowledge base will belong to. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled for + retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize your knowledge base. + ], + "uuid": "str" # Optional. Knowledge base id. + } + + # response body for status code(s): 200 + response == { + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + def update_knowledge_base( + self, + uuid: str, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Update a Knowledge Base. + + To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace + def update_knowledge_base( + self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Update a Knowledge Base. + + To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "database_id": "str", # Optional. The id of the DigitalOcean database this + knowledge base will use, optional. + "name": "str", # Optional. Knowledge base name. + "project_id": "str", # Optional. The id of the DigitalOcean project this + knowledge base will belong to. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled for + retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize your knowledge base. + ], + "uuid": "str" # Optional. Knowledge base id. + } + + # response body for status code(s): 200 + response == { + "knowledge_base": { + "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when + the knowledge base was added to the agent. + "created_at": "2020-02-20 00:00:00", # Optional. Creation date / + time. + "database_id": "str", # Optional. Knowledgebase Description. + "embedding_model_uuid": "str", # Optional. Knowledgebase + Description. + "is_public": bool, # Optional. Whether the knowledge base is public + or not. + "last_indexing_job": { + "completed_datasources": 0, # Optional. Number of + datasources indexed completed. + "created_at": "2020-02-20 00:00:00", # Optional. Creation + date / time. + "data_source_jobs": [ + { + "completed_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source completed indexing. + "data_source_uuid": "str", # Optional. Uuid + of the indexed data source. + "error_details": "str", # Optional. A + detailed error description. + "error_msg": "str", # Optional. A string + code provinding a hint which part of the system experienced an + error. + "failed_item_count": "str", # Optional. + Total count of files that have failed. + "indexed_file_count": "str", # Optional. + Total count of files that have been indexed. + "indexed_item_count": "str", # Optional. + Total count of files that have been indexed. + "removed_item_count": "str", # Optional. + Total count of files that have been removed. + "skipped_item_count": "str", # Optional. + Total count of files that have been skipped. + "started_at": "2020-02-20 00:00:00", # + Optional. Timestamp when data source started indexing. + "status": "DATA_SOURCE_STATUS_UNKNOWN", # + Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known + values are: "DATA_SOURCE_STATUS_UNKNOWN", + "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", + "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", + "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", + and "DATA_SOURCE_STATUS_CANCELLED". + "total_bytes": "str", # Optional. Total size + of files in data source in bytes. + "total_bytes_indexed": "str", # Optional. + Total size of files in data source in bytes that have been + indexed. + "total_file_count": "str" # Optional. Total + file count in the data source. + } + ], + "data_source_uuids": [ + "str" # Optional. IndexingJob description. + ], + "finished_at": "2020-02-20 00:00:00", # Optional. + IndexingJob description. + "is_report_available": bool, # Optional. Boolean value to + determine if the indexing job details are available. + "knowledge_base_uuid": "str", # Optional. Knowledge base id. + "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default + value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: + "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", + "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", + "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and + "BATCH_JOB_PHASE_CANCELLED". + "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob + description. + "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default + value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: + "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", + "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", + "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", + "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". + "tokens": 0, # Optional. Number of tokens [This field is + deprecated]. + "total_datasources": 0, # Optional. Number of datasources + being indexed. + "total_tokens": "str", # Optional. Total Tokens Consumed By + the Indexing Job. + "updated_at": "2020-02-20 00:00:00", # Optional. Last + modified. + "uuid": "str" # Optional. Unique id. + }, + "name": "str", # Optional. Name of knowledge base. + "project_id": "str", # Optional. Knowledgebase Description. + "region": "str", # Optional. Region code. + "reranking_config": { + "enabled": bool, # Optional. Whether reranking is enabled + for retrieval. + "model": "str" # Optional. Reranker model internal name. + }, + "tags": [ + "str" # Optional. Tags to organize related resources. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. + "user_id": "str", # Optional. Id of user that created the knowledge + base. + "uuid": "str" # Optional. Unique id for knowledge base. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) + cls: ClsType[JSON] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_update_knowledge_base_request( + uuid=uuid, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Delete a Knowledge Base. + + To delete a knowledge base, send a DELETE request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + + :param uuid: Knowledge base id. Required. + :type uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "uuid": "str" # Optional. The id of the deleted knowledge base. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_delete_knowledge_base_request( + uuid=uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @overload + def create_model_eval_dataset_upload_presigned_urls( # pylint: disable=name-too-long + self, + body: Optional[JSON] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Model Evaluation Dataset File Upload. + + To create presigned URLs for model evaluation dataset file upload, send a POST request to + ``/v2/genai/model_evaluation/datasets/file_upload_presigned_urls``. + + :param body: Default value is None. + :type body: JSON + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + def create_model_eval_dataset_upload_presigned_urls( # pylint: disable=name-too-long + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Create Presigned URLs for Model Evaluation Dataset File Upload. + + To create presigned URLs for model evaluation dataset file upload, send a POST request to + ``/v2/genai/model_evaluation/datasets/file_upload_presigned_urls``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace + def create_model_eval_dataset_upload_presigned_urls( # pylint: disable=name-too-long + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Delete a Data Source from a Knowledge Base. + """Create Presigned URLs for Model Evaluation Dataset File Upload. - To delete a data source from a knowledge base, send a DELETE request to - ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/data_sources/{data_source_uuid}``. + To create presigned URLs for model evaluation dataset file upload, send a POST request to + ``/v2/genai/model_evaluation/datasets/file_upload_presigned_urls``. - :param knowledge_base_uuid: Knowledge base id. Required. - :type knowledge_base_uuid: str - :param data_source_uuid: Data source id. Required. - :type data_source_uuid: str + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -250640,10 +263621,33 @@ def delete_knowledge_base_data_source( Example: .. code-block:: python + # JSON input template you can fill out and use as your body input. + body = { + "files": [ + { + "file_name": "str", # Optional. Local filename. + "file_size": "str" # Optional. The size of the file in + bytes. + } + ] + } + # response body for status code(s): 200 response == { - "data_source_uuid": "str", # Optional. Data source id. - "knowledge_base_uuid": "str" # Optional. Knowledge base id. + "request_id": "str", # Optional. The ID generated for the request for + Presigned URLs. + "uploads": [ + { + "expires_at": "2020-02-20 00:00:00", # Optional. The time + the url expires at. + "object_key": "str", # Optional. The unique object key to + store the file as. + "original_file_name": "str", # Optional. The original file + name. + "presigned_url": "str" # Optional. The actual presigned URL + the client can use to upload the file directly. + } + ] } # response body for status code(s): 404 response == { @@ -250670,14 +263674,29 @@ def delete_knowledge_base_data_source( } error_map.update(kwargs.pop("error_map", {}) or {}) - _headers = kwargs.pop("headers", {}) or {} + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} + content_type: Optional[str] = kwargs.pop( + "content_type", _headers.pop("Content-Type", None) + ) cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_delete_knowledge_base_data_source_request( - knowledge_base_uuid=knowledge_base_uuid, - data_source_uuid=data_source_uuid, + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(body, (IOBase, bytes)): + _content = body + else: + if body is not None: + _json = body + else: + _json = None + + _request = build_genai_create_model_eval_dataset_upload_presigned_urls_request( + content_type=content_type, + json=_json, + content=_content, headers=_headers, params=_params, ) @@ -250737,17 +263756,13 @@ def delete_knowledge_base_data_source( return cast(JSON, deserialized) # type: ignore @distributed_trace - def list_indexing_jobs_by_knowledge_base( - self, knowledge_base_uuid: str, **kwargs: Any - ) -> JSON: + def list_model_evaluation_metrics(self, **kwargs: Any) -> JSON: # pylint: disable=line-too-long - """List Indexing Jobs for a Knowledge Base. + """List Model Evaluation Metrics. - To list latest 15 indexing jobs for a knowledge base, send a GET request to - ``/v2/gen-ai/knowledge_bases/{knowledge_base_uuid}/indexing_jobs``. + To list all available metrics for model evaluation, send a GET request to + ``/v2/genai/model_evaluation_metrics``. - :param knowledge_base_uuid: Knowledge base uuid in string. Required. - :type knowledge_base_uuid: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -250757,97 +263772,47 @@ def list_indexing_jobs_by_knowledge_base( # response body for status code(s): 200 response == { - "jobs": [ + "metrics": [ { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. The indexing jobs. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. The - indexing jobs. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - } - ], - "links": { - "pages": { - "first": "str", # Optional. First page. - "last": "str", # Optional. Last page. - "next": "str", # Optional. Next page. - "previous": "str" # Optional. Previous page. + "category": "METRIC_CATEGORY_UNSPECIFIED", # Optional. + Default value is "METRIC_CATEGORY_UNSPECIFIED". Known values are: + "METRIC_CATEGORY_UNSPECIFIED", "METRIC_CATEGORY_CORRECTNESS", + "METRIC_CATEGORY_USER_OUTCOMES", "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. List of model evaluation + metrics. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", # + Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. List of model evaluation + metrics. + "metric_name": "str", # Optional. List of model evaluation + metrics. + "metric_rank": 0, # Optional. List of model evaluation + metrics. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. + Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. List of model evaluation + metrics. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known values + are: "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for the + metric. + "range_min": 0.0 # Optional. The minimum value for the + metric. } - }, - "meta": { - "page": 0, # Optional. The current page. - "pages": 0, # Optional. Total number of pages. - "total": 0 # Optional. Total amount of items over all pages. - } + ] } # response body for status code(s): 404 response == { @@ -250879,8 +263844,7 @@ def list_indexing_jobs_by_knowledge_base( cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_list_indexing_jobs_by_knowledge_base_request( - knowledge_base_uuid=knowledge_base_uuid, + _request = build_genai_list_model_evaluation_metrics_request( headers=_headers, params=_params, ) @@ -250940,15 +263904,33 @@ def list_indexing_jobs_by_knowledge_base( return cast(JSON, deserialized) # type: ignore @distributed_trace - def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + def list_model_evaluation_runs( + self, + *, + eval_preset_uuid: Optional[str] = None, + status: str = "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any, + ) -> JSON: # pylint: disable=line-too-long - """Retrieve Information About an Existing Knowledge Base. - - To retrive information about an existing knowledge base, send a GET request to - ``/v2/gen-ai/knowledge_bases/{uuid}``. - - :param uuid: Knowledge base id. Required. - :type uuid: str + """List Model Evaluation Runs. + + To list model evaluation runs, send a GET request to ``/v2/genai/model_evaluation_runs``. + + :keyword eval_preset_uuid: UUID of the evaluation preset to filter by. Default value is None. + :paramtype eval_preset_uuid: str + :keyword status: Filter by evaluation run status. Known values are: + "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", "MODEL_EVALUATION_RUN_QUEUED", + "MODEL_EVALUATION_RUN_RUNNING_DATASET", "MODEL_EVALUATION_RUN_EVALUATING_RESULTS", + "MODEL_EVALUATION_RUN_CANCELLING", "MODEL_EVALUATION_RUN_CANCELLED", + "MODEL_EVALUATION_RUN_SUCCESSFUL", "MODEL_EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "MODEL_EVALUATION_RUN_FAILED". Default value is "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED". + :paramtype status: str + :keyword page: Page number. Default value is None. + :paramtype page: int + :keyword per_page: Items per page. Default value is None. + :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -250958,112 +263940,56 @@ def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "database_status": "CREATING", # Optional. Default value is "CREATING". - Known values are: "CREATING", "ONLINE", "POWEROFF", "REBUILDING", "REBALANCING", - "DECOMMISSIONED", "FORKING", "MIGRATING", "RESIZING", "RESTORING", "POWERING_ON", - and "UNHEALTHY". - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "runs": [ + { + "candidate_model_name": "str", # Optional. Name of the + candidate model being evaluated. + "candidate_model_source": + "CANDIDATE_MODEL_SOURCE_SERVERLESS", # Optional. Default value is + "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether inference runs against the + serverless platform, a dedicated deployment, or a model router. Known + values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. UUID of the + candidate model being evaluated. + "created_at": "2020-02-20 00:00:00", # Optional. Timestamp + when the run was created. + "dataset_name": "str", # Optional. Name of the dataset used + for evaluation. + "dataset_uuid": "str", # Optional. UUID of the dataset used + for evaluation. + "eval_run_uuid": "str", # Optional. UUID of the evaluation + run. + "judge_model_name": "str", # Optional. Summary view of + evaluation runs for the run history list. + "judge_model_uuid": "str", # Optional. Judge model used to + score responses. + "name": "str", # Optional. Name of the evaluation run. + "status": "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED" # + Optional. Default value is "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED". + Model Evaluation Run Statuses. Known values are: + "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", "MODEL_EVALUATION_RUN_QUEUED", + "MODEL_EVALUATION_RUN_RUNNING_DATASET", + "MODEL_EVALUATION_RUN_EVALUATING_RESULTS", + "MODEL_EVALUATION_RUN_CANCELLING", "MODEL_EVALUATION_RUN_CANCELLED", + "MODEL_EVALUATION_RUN_SUCCESSFUL", + "MODEL_EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "MODEL_EVALUATION_RUN_FAILED". + } + ] } # response body for status code(s): 404 response == { @@ -251095,8 +264021,11 @@ def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_get_knowledge_base_request( - uuid=uuid, + _request = build_genai_list_model_evaluation_runs_request( + eval_preset_uuid=eval_preset_uuid, + status=status, + page=page, + per_page=per_page, headers=_headers, params=_params, ) @@ -251156,21 +264085,18 @@ def get_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @overload - def update_knowledge_base( + def create_model_evaluation_run( self, - uuid: str, body: Optional[JSON] = None, *, content_type: str = "application/json", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Update a Knowledge Base. + """Create Model Evaluation Run. - To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To create a model evaluation run, send a POST request to ``/v2/genai/model_evaluation_runs``. - :param uuid: Knowledge base id. Required. - :type uuid: str :param body: Default value is None. :type body: JSON :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. @@ -251185,128 +264111,55 @@ def update_knowledge_base( # JSON input template you can fill out and use as your body input. body = { - "database_id": "str", # Optional. The id of the DigitalOcean database this - knowledge base will use, optiona. - "embedding_model_uuid": "str", # Optional. Identifier for the foundation - model. - "name": "str", # Optional. Knowledge base name. - "project_id": "str", # Optional. The id of the DigitalOcean project this - knowledge base will belong to. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled for - retrieval. - "model": "str" # Optional. Reranker model internal name. + "candidate_inference_config": { + "max_tokens": 0, # Optional. Inference configuration for the + candidate model during evaluation. + "stop_token": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "system_prompt": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "temperature": 0.0 # Optional. Inference configuration for the + candidate model during evaluation. }, - "tags": [ - "str" # Optional. Tags to organize your knowledge base. + "candidate_model_name": "str", # Optional. Model slug used to call the + candidate model API. For dedicated inference, this is the model slug from the + deployment. For serverless, this should match the model's internal name. + "candidate_model_source": "CANDIDATE_MODEL_SOURCE_SERVERLESS", # Optional. + Default value is "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether inference runs + against the serverless platform, a dedicated deployment, or a model router. Known + values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. UUID of the candidate model to + evaluate. + "dataset_uuid": "str", # Optional. UUID of the dataset to use for + evaluation. + "eval_preset_uuid": "str", # Optional. + "judge_model_uuid": "str", # Optional. UUID of the judge model used to score + responses. + "metric_uuids": [ + "str" # Optional. UUIDs of metrics to evaluate (selected from + ListModelEvaluationMetrics). ], - "uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. + "preset_name": "str", # Optional. + "save_as_preset": bool, # Optional. .. role:: raw-html-m2r(raw) :format: + html If true, saves the inline config as a reusable preset"" + :raw-html-m2r:`
` Ignored when eval_preset_uuid is provided. + "source": "str", # Optional. Source of the run creation (api, sdk, cli). + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + } } # response body for status code(s): 200 response == { - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "eval_run_uuid": "str" # Optional. UUID of the created evaluation run. } # response body for status code(s): 404 response == { @@ -251322,21 +264175,18 @@ def update_knowledge_base( """ @overload - def update_knowledge_base( + def create_model_evaluation_run( self, - uuid: str, body: Optional[IO[bytes]] = None, *, content_type: str = "application/json", **kwargs: Any, ) -> JSON: # pylint: disable=line-too-long - """Update a Knowledge Base. + """Create Model Evaluation Run. - To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To create a model evaluation run, send a POST request to ``/v2/genai/model_evaluation_runs``. - :param uuid: Knowledge base id. Required. - :type uuid: str :param body: Default value is None. :type body: IO[bytes] :keyword content_type: Body Parameter content-type. Content type parameter for binary body. @@ -251351,108 +264201,7 @@ def update_knowledge_base( # response body for status code(s): 200 response == { - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "eval_run_uuid": "str" # Optional. UUID of the created evaluation run. } # response body for status code(s): 404 response == { @@ -251468,16 +264217,14 @@ def update_knowledge_base( """ @distributed_trace - def update_knowledge_base( - self, uuid: str, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + def create_model_evaluation_run( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any ) -> JSON: # pylint: disable=line-too-long - """Update a Knowledge Base. + """Create Model Evaluation Run. - To update a knowledge base, send a PUT request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To create a model evaluation run, send a POST request to ``/v2/genai/model_evaluation_runs``. - :param uuid: Knowledge base id. Required. - :type uuid: str :param body: Is either a JSON type or a IO[bytes] type. Default value is None. :type body: JSON or IO[bytes] :return: JSON object @@ -251489,128 +264236,55 @@ def update_knowledge_base( # JSON input template you can fill out and use as your body input. body = { - "database_id": "str", # Optional. The id of the DigitalOcean database this - knowledge base will use, optiona. - "embedding_model_uuid": "str", # Optional. Identifier for the foundation - model. - "name": "str", # Optional. Knowledge base name. - "project_id": "str", # Optional. The id of the DigitalOcean project this - knowledge base will belong to. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled for - retrieval. - "model": "str" # Optional. Reranker model internal name. + "candidate_inference_config": { + "max_tokens": 0, # Optional. Inference configuration for the + candidate model during evaluation. + "stop_token": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "system_prompt": "str", # Optional. Inference configuration for the + candidate model during evaluation. + "temperature": 0.0 # Optional. Inference configuration for the + candidate model during evaluation. }, - "tags": [ - "str" # Optional. Tags to organize your knowledge base. + "candidate_model_name": "str", # Optional. Model slug used to call the + candidate model API. For dedicated inference, this is the model slug from the + deployment. For serverless, this should match the model's internal name. + "candidate_model_source": "CANDIDATE_MODEL_SOURCE_SERVERLESS", # Optional. + Default value is "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether inference runs + against the serverless platform, a dedicated deployment, or a model router. Known + values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. UUID of the candidate model to + evaluate. + "dataset_uuid": "str", # Optional. UUID of the dataset to use for + evaluation. + "eval_preset_uuid": "str", # Optional. + "judge_model_uuid": "str", # Optional. UUID of the judge model used to score + responses. + "metric_uuids": [ + "str" # Optional. UUIDs of metrics to evaluate (selected from + ListModelEvaluationMetrics). ], - "uuid": "str" # Optional. Knowledge base id. + "name": "str", # Optional. + "preset_name": "str", # Optional. + "save_as_preset": bool, # Optional. .. role:: raw-html-m2r(raw) :format: + html If true, saves the inline config as a reusable preset"" + :raw-html-m2r:`
` Ignored when eval_preset_uuid is provided. + "source": "str", # Optional. Source of the run creation (api, sdk, cli). + "star_metric": { + "metric_uuid": "str", # Optional. + "name": "str", # Optional. + "success_threshold": 0.0, # Optional. The success threshold for the + star metric. This is a value that the metric must reach to be considered + successful. + "success_threshold_pct": 0 # Optional. The success threshold for the + star metric. This is a percentage value between 0 and 100. + } } # response body for status code(s): 200 response == { - "knowledge_base": { - "added_to_agent_at": "2020-02-20 00:00:00", # Optional. Time when - the knowledge base was added to the agent. - "created_at": "2020-02-20 00:00:00", # Optional. Creation date / - time. - "database_id": "str", # Optional. Knowledgebase Description. - "embedding_model_uuid": "str", # Optional. Knowledgebase - Description. - "is_public": bool, # Optional. Whether the knowledge base is public - or not. - "last_indexing_job": { - "completed_datasources": 0, # Optional. Number of - datasources indexed completed. - "created_at": "2020-02-20 00:00:00", # Optional. Creation - date / time. - "data_source_jobs": [ - { - "completed_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source completed indexing. - "data_source_uuid": "str", # Optional. Uuid - of the indexed data source. - "error_details": "str", # Optional. A - detailed error description. - "error_msg": "str", # Optional. A string - code provinding a hint which part of the system experienced an - error. - "failed_item_count": "str", # Optional. - Total count of files that have failed. - "indexed_file_count": "str", # Optional. - Total count of files that have been indexed. - "indexed_item_count": "str", # Optional. - Total count of files that have been indexed. - "removed_item_count": "str", # Optional. - Total count of files that have been removed. - "skipped_item_count": "str", # Optional. - Total count of files that have been skipped. - "started_at": "2020-02-20 00:00:00", # - Optional. Timestamp when data source started indexing. - "status": "DATA_SOURCE_STATUS_UNKNOWN", # - Optional. Default value is "DATA_SOURCE_STATUS_UNKNOWN". Known - values are: "DATA_SOURCE_STATUS_UNKNOWN", - "DATA_SOURCE_STATUS_IN_PROGRESS", "DATA_SOURCE_STATUS_UPDATED", - "DATA_SOURCE_STATUS_PARTIALLY_UPDATED", - "DATA_SOURCE_STATUS_NOT_UPDATED", "DATA_SOURCE_STATUS_FAILED", - and "DATA_SOURCE_STATUS_CANCELLED". - "total_bytes": "str", # Optional. Total size - of files in data source in bytes. - "total_bytes_indexed": "str", # Optional. - Total size of files in data source in bytes that have been - indexed. - "total_file_count": "str" # Optional. Total - file count in the data source. - } - ], - "data_source_uuids": [ - "str" # Optional. IndexingJob description. - ], - "finished_at": "2020-02-20 00:00:00", # Optional. - IndexingJob description. - "is_report_available": bool, # Optional. Boolean value to - determine if the indexing job details are available. - "knowledge_base_uuid": "str", # Optional. Knowledge base id. - "phase": "BATCH_JOB_PHASE_UNKNOWN", # Optional. Default - value is "BATCH_JOB_PHASE_UNKNOWN". Known values are: - "BATCH_JOB_PHASE_UNKNOWN", "BATCH_JOB_PHASE_PENDING", - "BATCH_JOB_PHASE_RUNNING", "BATCH_JOB_PHASE_SUCCEEDED", - "BATCH_JOB_PHASE_FAILED", "BATCH_JOB_PHASE_ERROR", and - "BATCH_JOB_PHASE_CANCELLED". - "started_at": "2020-02-20 00:00:00", # Optional. IndexingJob - description. - "status": "INDEX_JOB_STATUS_UNKNOWN", # Optional. Default - value is "INDEX_JOB_STATUS_UNKNOWN". Known values are: - "INDEX_JOB_STATUS_UNKNOWN", "INDEX_JOB_STATUS_PARTIAL", - "INDEX_JOB_STATUS_IN_PROGRESS", "INDEX_JOB_STATUS_COMPLETED", - "INDEX_JOB_STATUS_FAILED", "INDEX_JOB_STATUS_NO_CHANGES", - "INDEX_JOB_STATUS_PENDING", and "INDEX_JOB_STATUS_CANCELLED". - "tokens": 0, # Optional. Number of tokens [This field is - deprecated]. - "total_datasources": 0, # Optional. Number of datasources - being indexed. - "total_tokens": "str", # Optional. Total Tokens Consumed By - the Indexing Job. - "updated_at": "2020-02-20 00:00:00", # Optional. Last - modified. - "uuid": "str" # Optional. Unique id. - }, - "name": "str", # Optional. Name of knowledge base. - "project_id": "str", # Optional. Knowledgebase Description. - "region": "str", # Optional. Region code. - "reranking_config": { - "enabled": bool, # Optional. Whether reranking is enabled - for retrieval. - "model": "str" # Optional. Reranker model internal name. - }, - "tags": [ - "str" # Optional. Tags to organize related resources. - ], - "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. - "user_id": "str", # Optional. Id of user that created the knowledge - base. - "uuid": "str" # Optional. Unique id for knowledge base. - } + "eval_run_uuid": "str" # Optional. UUID of the created evaluation run. } # response body for status code(s): 404 response == { @@ -251656,8 +264330,7 @@ def update_knowledge_base( else: _json = None - _request = build_genai_update_knowledge_base_request( - uuid=uuid, + _request = build_genai_create_model_evaluation_run_request( content_type=content_type, json=_json, content=_content, @@ -251720,14 +264393,27 @@ def update_knowledge_base( return cast(JSON, deserialized) # type: ignore @distributed_trace - def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: + def get_model_evaluation_run( + self, + eval_run_uuid: str, + *, + page: Optional[int] = None, + per_page: Optional[int] = None, + **kwargs: Any, + ) -> JSON: # pylint: disable=line-too-long - """Delete a Knowledge Base. + """Retrieve Model Evaluation Run. - To delete a knowledge base, send a DELETE request to ``/v2/gen-ai/knowledge_bases/{uuid}``. + To retrieve a model evaluation run, send a GET request to + ``/v2/genai/model_evaluation_runs/{eval_run_uuid}``. - :param uuid: Knowledge base id. Required. - :type uuid: str + :param eval_run_uuid: UUID of the evaluation run. Required. + :type eval_run_uuid: str + :keyword page: Page number for per-prompt results (defaults to 1). Default value is None. + :paramtype page: int + :keyword per_page: Number of per-prompt results per page (defaults to 50). Default value is + None. + :paramtype per_page: int :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -251737,7 +264423,323 @@ def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "uuid": "str" # Optional. The id of the deleted knowledge base. + "links": { + "pages": { + "first": "str", # Optional. First page. + "last": "str", # Optional. Last page. + "next": "str", # Optional. Next page. + "previous": "str" # Optional. Previous page. + } + }, + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + }, + "results": [ + { + "candidate_model_name": "str", # Optional. Paginated + per-prompt evaluation results. + "candidate_model_uuid": "str", # Optional. Paginated + per-prompt evaluation results. + "ground_truth": "str", # Optional. Paginated per-prompt + evaluation results. + "input": "str", # Optional. The input query sent to the + candidate model. + "metric_results": [ + { + "error_description": "str", # Optional. + Error description if the metric could not be calculated. + "metric_name": "str", # Optional. Metric + name. + "metric_value_type": + "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is + "METRIC_VALUE_TYPE_UNSPECIFIED". Known values are: + "METRIC_VALUE_TYPE_UNSPECIFIED", "METRIC_VALUE_TYPE_NUMBER", + "METRIC_VALUE_TYPE_STRING", and "METRIC_VALUE_TYPE_PERCENTAGE". + "number_value": 0.0, # Optional. The value + of the metric as a number. + "reasoning": "str", # Optional. Reasoning of + the metric result. + "string_value": "str" # Optional. The value + of the metric as a string. + } + ], + "output": "str" # Optional. The response from the candidate + model. + } + ], + "run": { + "candidate_inference_config": { + "max_tokens": 0, # Optional. Inference configuration for the + candidate model during evaluation. + "stop_token": "str", # Optional. Inference configuration for + the candidate model during evaluation. + "system_prompt": "str", # Optional. Inference configuration + for the candidate model during evaluation. + "temperature": 0.0 # Optional. Inference configuration for + the candidate model during evaluation. + }, + "candidate_model_name": "str", # Optional. Model Evaluation Run + Detail - full view returned when fetching a specific run. + "candidate_model_source": "CANDIDATE_MODEL_SOURCE_SERVERLESS", # + Optional. Default value is "CANDIDATE_MODEL_SOURCE_SERVERLESS". Whether + inference runs against the serverless platform, a dedicated deployment, or a + model router. Known values are: "CANDIDATE_MODEL_SOURCE_SERVERLESS", + "CANDIDATE_MODEL_SOURCE_DEDICATED", and "CANDIDATE_MODEL_SOURCE_ROUTER". + "candidate_model_uuid": "str", # Optional. Candidate model being + evaluated. + "completed_at": "2020-02-20 00:00:00", # Optional. Model Evaluation + Run Detail - full view returned when fetching a specific run. + "created_at": "2020-02-20 00:00:00", # Optional. Model Evaluation + Run Detail - full view returned when fetching a specific run. + "dataset_name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "dataset_uuid": "str", # Optional. Dataset used for the evaluation. + "error_description": "str", # Optional. Error description if the run + failed or partially succeeded. + "eval_preset_name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "eval_preset_uuid": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "eval_run_uuid": "str", # Optional. UUID of the evaluation run. + "judge_model_name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "judge_model_uuid": "str", # Optional. Judge model used to score + responses. + "metrics": [ + { + "category": "METRIC_CATEGORY_UNSPECIFIED", # + Optional. Default value is "METRIC_CATEGORY_UNSPECIFIED". Known + values are: "METRIC_CATEGORY_UNSPECIFIED", + "METRIC_CATEGORY_CORRECTNESS", "METRIC_CATEGORY_USER_OUTCOMES", + "METRIC_CATEGORY_SAFETY_AND_SECURITY", + "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". + "description": "str", # Optional. Metrics selected + for this evaluation. + "evaluation_scope": "EVALUATION_SCOPE_UNSPECIFIED", + # Optional. Default value is "EVALUATION_SCOPE_UNSPECIFIED". Scope + that determines whether a metric belongs to agent evaluation or model + evaluation. For backwards compatibility, UNSPECIFIED defaults to + agent metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". + "inverted": bool, # Optional. If true, the metric is + inverted, meaning that a lower value is better. + "is_metric_goal": bool, # Optional. Metrics selected + for this evaluation. + "metric_name": "str", # Optional. Metrics selected + for this evaluation. + "metric_rank": 0, # Optional. Metrics selected for + this evaluation. + "metric_type": "METRIC_TYPE_UNSPECIFIED", # + Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values + are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". + "metric_uuid": "str", # Optional. Metrics selected + for this evaluation. + "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", + # Optional. Default value is "METRIC_VALUE_TYPE_UNSPECIFIED". Known + values are: "METRIC_VALUE_TYPE_UNSPECIFIED", + "METRIC_VALUE_TYPE_NUMBER", "METRIC_VALUE_TYPE_STRING", and + "METRIC_VALUE_TYPE_PERCENTAGE". + "range_max": 0.0, # Optional. The maximum value for + the metric. + "range_min": 0.0 # Optional. The minimum value for + the metric. + } + ], + "name": "str", # Optional. Name of the evaluation run. + "result_summary": { + "end_time": "2020-02-20 00:00:00", # Optional. Aggregated + result summary for a completed model evaluation run. + "metric_summaries": [ + { + "description": "str", # Optional. Per-metric + aggregated pass/fail statistics. + "fail_percent": 0.0, # Optional. Per-metric + aggregated pass/fail statistics. + "metric_name": "str", # Optional. Per-metric + aggregated pass/fail statistics. + "metric_uuid": "str", # Optional. Per-metric + aggregated pass/fail statistics. + "pass_percent": 0.0 # Optional. Per-metric + aggregated pass/fail statistics. + } + ], + "overall_score_percent": 0.0, # Optional. Aggregated result + summary for a completed model evaluation run. + "per_model_summaries": { + "summaries": [ + { + "metric_summaries": [ + { + "description": "str", + # Optional. Per-metric pass/fail for only this + model's prompts. + "fail_percent": 0.0, + # Optional. Per-metric pass/fail for only this + model's prompts. + "metric_name": "str", + # Optional. Per-metric pass/fail for only this + model's prompts. + "metric_uuid": "str", + # Optional. Per-metric pass/fail for only this + model's prompts. + "pass_percent": 0.0 + # Optional. Per-metric pass/fail for only this + model's prompts. + } + ], + "model_name": "str", # Optional. + Name/slug of the model (matches routed_model from results). + "performance_metrics": { + "candidate_latency": { + "avg_e2e_latency_ms": + 0.0, # Optional. Average end-to-end latency across + all invocations. + "max_e2e_latency_ms": + 0.0, # Optional. Maximum end-to-end latency + observed. + "min_e2e_latency_ms": + 0.0, # Optional. Minimum end-to-end latency + observed. + "p50_latency_ms": + 0.0, # Optional. P50 (median) latency. + "p90_latency_ms": + 0.0, # Optional. P90 latency. + "p95_latency_ms": 0.0 + # Optional. P95 latency. + }, + "token_usage": { + "total_candidate_input_tokens": "str", # Optional. + All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_output_tokens": "str", # Optional. + All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_tokens": "str", # Optional. All + performance metrics are for the candidate model + unless noted otherwise. + "total_judge_input_tokens": "str", # Optional. All + performance metrics are for the candidate model + unless noted otherwise. + "total_judge_output_tokens": "str", # Optional. All + performance metrics are for the candidate model + unless noted otherwise. + "total_judge_tokens": + "str" # Optional. All performance metrics are for + the candidate model unless noted otherwise. + } + }, + "prompt_count": 0 # Optional. Number + of prompts routed to this model. + } + ] + }, + "performance_metrics": { + "candidate_latency": { + "avg_e2e_latency_ms": 0.0, # Optional. + Average end-to-end latency across all invocations. + "max_e2e_latency_ms": 0.0, # Optional. + Maximum end-to-end latency observed. + "min_e2e_latency_ms": 0.0, # Optional. + Minimum end-to-end latency observed. + "p50_latency_ms": 0.0, # Optional. P50 + (median) latency. + "p90_latency_ms": 0.0, # Optional. P90 + latency. + "p95_latency_ms": 0.0 # Optional. P95 + latency. + }, + "token_usage": { + "total_candidate_input_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_output_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_candidate_tokens": "str", # Optional. + All performance metrics are for the candidate model unless noted + otherwise. + "total_judge_input_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_judge_output_tokens": "str", # + Optional. All performance metrics are for the candidate model + unless noted otherwise. + "total_judge_tokens": "str" # Optional. All + performance metrics are for the candidate model unless noted + otherwise. + } + }, + "pricing": { + "currency": "str", # Optional. Currency code (e.g., + "USD"). + "judge_model_pricing": { + "input_cost": 0.0, # Optional. Cost of input + tokens. + "output_cost": 0.0, # Optional. Cost of + output tokens. + "total_cost": 0.0 # Optional. Total cost + (input + output). + }, + "per_candidate_model_pricing": [ + { + "model_name": "str", # Optional. + Model name (for display purposes). + "model_uuid": "str", # Optional. + Model UUID. + "pricing": { + "input_cost": 0.0, # + Optional. Cost of input tokens. + "output_cost": 0.0, # + Optional. Cost of output tokens. + "total_cost": 0.0 # + Optional. Total cost (input + output). + }, + "prompt_count": 0 # Optional. Number + of prompts/rows routed to this model. + } + ], + "total_cost": 0.0 # Optional. Total cost of the + evaluation run (all candidates + judge). + }, + "star_metric_summary": { + "metric_name": "str", # Optional. Star metric + summary with identifying details and threshold. + "metric_uuid": "str", # Optional. Star metric + summary with identifying details and threshold. + "threshold": 0.0 # Optional. Star metric summary + with identifying details and threshold. + }, + "start_time": "2020-02-20 00:00:00", # Optional. Run timing. + "total_duration_seconds": 0 # Optional. Total wall-clock + duration in seconds. + }, + "star_metric": { + "metric_uuid": "str", # Optional. Model Evaluation Run + Detail - full view returned when fetching a specific run. + "name": "str", # Optional. Model Evaluation Run Detail - + full view returned when fetching a specific run. + "success_threshold": 0.0, # Optional. The success threshold + for the star metric. This is a value that the metric must reach to be + considered successful. + "success_threshold_pct": 0 # Optional. The success threshold + for the star metric. This is a percentage value between 0 and 100. + }, + "started_at": "2020-02-20 00:00:00", # Optional. Model Evaluation + Run Detail - full view returned when fetching a specific run. + "status": "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED" # Optional. + Default value is "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED". Model Evaluation + Run Statuses. Known values are: "MODEL_EVALUATION_RUN_STATUS_UNSPECIFIED", + "MODEL_EVALUATION_RUN_QUEUED", "MODEL_EVALUATION_RUN_RUNNING_DATASET", + "MODEL_EVALUATION_RUN_EVALUATING_RESULTS", "MODEL_EVALUATION_RUN_CANCELLING", + "MODEL_EVALUATION_RUN_CANCELLED", "MODEL_EVALUATION_RUN_SUCCESSFUL", + "MODEL_EVALUATION_RUN_PARTIALLY_SUCCESSFUL", and + "MODEL_EVALUATION_RUN_FAILED". + } } # response body for status code(s): 404 response == { @@ -251769,8 +264771,125 @@ def delete_knowledge_base(self, uuid: str, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_delete_knowledge_base_request( - uuid=uuid, + _request = build_genai_get_model_evaluation_run_request( + eval_run_uuid=eval_run_uuid, + page=page, + per_page=per_page, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def get_model_evaluation_run_results_download_url( # pylint: disable=name-too-long + self, eval_run_uuid: str, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Get Download URL for Model Evaluation Run Results. + + To get a presigned download URL for model evaluation run results (gzip-compressed JSON), send a + GET request to ``/v2/genai/model_evaluation_runs/{eval_run_uuid}/results/download_url``. + + :param eval_run_uuid: UUID of the evaluation run. Required. + :type eval_run_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "download_url": "str", # Optional. The presigned URL to download the + gzip-compressed JSON results file (.json.gz). + "expires_at": "2020-02-20 00:00:00" # Optional. The time the URL expires at. + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_get_model_evaluation_run_results_download_url_request( + eval_run_uuid=eval_run_uuid, headers=_headers, params=_params, ) @@ -251892,8 +265011,28 @@ def list_models( "url": "str", # Optional. Agreement Description. "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. Model capabilities (inference, + reasoning, vectorization, etc.). + ], + "context_window": "str", # Optional. Context window (maximum + tokens). "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "description": "str", # Optional. Model description. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities + supported by this endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "id": "str", # Optional. Human-readable model identifier. "is_foundational": bool, # Optional. True if it is a foundational model provided by do. @@ -251903,9 +265042,79 @@ def list_models( of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle status of + the model (internal, public-preview, active, deprecated, end_of_life). + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, + "model_availability": "str", # Optional. Model availability + (serverless, dedicated, etc.). "name": "str", # Optional. Display name of the model. + "parameter_count": 0.0, # Optional. Parameter count in + billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. + "pricing": { + "input_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "output_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "price_per_audio": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_image": 0.0, # Optional. Unit-based + pricing for non-token models (e.g., Fal AI image/video/audio + generation, speech models). At most one of these is typically + populated per model. Token-based models (chat, embeddings) leave all + of these at 0 and populate input_price_per_million / + output_price_per_million instead. + "price_per_megapixel": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_second": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_thousand_characters": 0.0, # Optional. + Pricing per million tokens (aligns with existing ModelPrice pattern). + "price_per_video": 0.0 # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + }, + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. + Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: + "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and + "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning efforts for + this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model supports + extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, embedding, + image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been fully @@ -252775,15 +265984,305 @@ def delete_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: return cast(JSON, deserialized) # type: ignore @distributed_trace - def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: + def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: + # pylint: disable=line-too-long + """Regenerate API Key for a Model. + + To regenerate a model API key, send a PUT request to + ``/v2/gen-ai/models/api_keys/{api_key_uuid}/regenerate``. + + :param api_key_uuid: API key ID. Required. + :type api_key_uuid: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "api_key_info": { + "created_at": "2020-02-20 00:00:00", # Optional. Creation date. + "created_by": "str", # Optional. Created by. + "deleted_at": "2020-02-20 00:00:00", # Optional. Deleted date. + "name": "str", # Optional. Name. + "secret_key": "str", # Optional. Model API Key Info. + "uuid": "str" # Optional. Uuid. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_regenerate_model_api_key_request( + api_key_uuid=api_key_uuid, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def list_model_catalog( + self, *, page: Optional[int] = None, limit: Optional[int] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """List Model Catalog. + + Returns all available models. + + :keyword page: Default value is None. + :paramtype page: int + :keyword limit: Default value is None. + :paramtype limit: int + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "data": [ + { + "availability": [ + "str" # Optional. + ], + "benchmark_score": {}, # Optional. Benchmark scores for this + model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. + ], + "context_window": "str", # Optional. Specs (flat). + "creator": "str", # Optional. Model creator/developer (e.g., + "Meta", "Anthropic", "OpenAI"). + "id": "str", # Optional. Identity. + "model_id": "str", # Optional. Model identifier used for API + calls (e.g., "llama3.1-70b-instruct"). + "name": "str", # Optional. + "parameter_count": 0.0, # Optional. + "pricing": { + "input_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "output_price_per_million": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "price_per_audio": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_image": 0.0, # Optional. Unit-based + pricing for non-token models (e.g., Fal AI image/video/audio + generation, speech models). At most one of these is typically + populated per model. Token-based models (chat, embeddings) leave all + of these at 0 and populate input_price_per_million / + output_price_per_million instead. + "price_per_megapixel": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_second": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_thousand_characters": 0.0, # Optional. + Pricing per million tokens (aligns with existing ModelPrice pattern). + "price_per_video": 0.0 # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + }, + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. + Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: + "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and + "MODEL_PROVIDER_OPENAI". + "short_description": "str", # Optional. + "type": "str" # Optional. + } + ], + "meta": { + "page": 0, # Optional. The current page. + "pages": 0, # Optional. Total number of pages. + "total": 0 # Optional. Total amount of items over all pages. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + error_map: MutableMapping[int, Type[HttpResponseError]] = { + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + 401: cast( + Type[HttpResponseError], + lambda response: ClientAuthenticationError(response=response), + ), + 429: HttpResponseError, + 500: HttpResponseError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[JSON] = kwargs.pop("cls", None) + + _request = build_genai_list_model_catalog_request( + page=page, + limit=limit, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = ( + self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 404]: + if _stream: + response.read() # Load the body in memory and close the socket + map_error(status_code=response.status_code, response=response, error_map=error_map) # type: ignore + raise HttpResponseError(response=response) + + response_headers = {} + if response.status_code == 200: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if response.status_code == 404: + response_headers["ratelimit-limit"] = self._deserialize( + "int", response.headers.get("ratelimit-limit") + ) + response_headers["ratelimit-remaining"] = self._deserialize( + "int", response.headers.get("ratelimit-remaining") + ) + response_headers["ratelimit-reset"] = self._deserialize( + "int", response.headers.get("ratelimit-reset") + ) + + if response.content: + deserialized = response.json() + else: + deserialized = None + + if cls: + return cls(pipeline_response, cast(JSON, deserialized), response_headers) # type: ignore + + return cast(JSON, deserialized) # type: ignore + + @distributed_trace + def get_model_catalog_card( + self, id: str, *, model_id: Optional[str] = None, **kwargs: Any + ) -> JSON: # pylint: disable=line-too-long - """Regenerate API Key for a Model. + """Get Model Catalog Card. - To regenerate a model API key, send a PUT request to - ``/v2/gen-ai/models/api_keys/{api_key_uuid}/regenerate``. + Returns detailed information for a specific model in the catalog including capabilities, + pricing, and code examples. - :param api_key_uuid: API key ID. Required. - :type api_key_uuid: str + :param id: Required. + :type id: str + :keyword model_id: Model identifier used for API calls (e.g., "llama3.1-70b-instruct"). + Alternative to UUID lookup. Default value is None. + :paramtype model_id: str :return: JSON object :rtype: JSON :raises ~azure.core.exceptions.HttpResponseError: @@ -252793,13 +266292,70 @@ def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: # response body for status code(s): 200 response == { - "api_key_info": { - "created_at": "2020-02-20 00:00:00", # Optional. Creation date. - "created_by": "str", # Optional. Created by. - "deleted_at": "2020-02-20 00:00:00", # Optional. Deleted date. - "name": "str", # Optional. Name. - "secret_key": "str", # Optional. Model API Key Info. - "uuid": "str" # Optional. Uuid. + "data": { + "availability": [ + "str" # Optional. Detail view for GetModelCatalogCard. + ], + "benchmark_score": {}, # Optional. Benchmark scores for this model, + stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. Detail view for GetModelCatalogCard. + ], + "code_snippets": { + "curl": "str", # Optional. Code examples for using the + model. + "javascript": "str", # Optional. Code examples for using the + model. + "python": "str", # Optional. Code examples for using the + model. + "sdk": "str" # Optional. Code examples for using the model. + }, + "context_window": "str", # Optional. Specs (same as Entry). + "creator": "str", # Optional. Model creator/developer (e.g., "Meta", + "Anthropic", "OpenAI"). + "description": "str", # Optional. Card-specific. + "id": "str", # Optional. Identity (same as Entry). + "modalities": { + "input": [ + "str" # Optional. Input/output modalities. + ], + "output": [ + "str" # Optional. Input/output modalities. + ] + }, + "model_id": "str", # Optional. Model identifier used for API calls + (e.g., "llama3.1-70b-instruct"). + "name": "str", # Optional. Detail view for GetModelCatalogCard. + "parameter_count": 0.0, # Optional. Detail view for + GetModelCatalogCard. + "pricing": { + "input_price_per_million": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "output_price_per_million": 0.0, # Optional. Pricing per + million tokens (aligns with existing ModelPrice pattern). + "price_per_audio": 0.0, # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + "price_per_image": 0.0, # Optional. Unit-based pricing for + non-token models (e.g., Fal AI image/video/audio generation, speech + models). At most one of these is typically populated per model. + Token-based models (chat, embeddings) leave all of these at 0 and + populate input_price_per_million / output_price_per_million instead. + "price_per_megapixel": 0.0, # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + "price_per_second": 0.0, # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + "price_per_thousand_characters": 0.0, # Optional. Pricing + per million tokens (aligns with existing ModelPrice pattern). + "price_per_video": 0.0 # Optional. Pricing per million + tokens (aligns with existing ModelPrice pattern). + }, + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value + is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: + "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and + "MODEL_PROVIDER_OPENAI". + "short_description": "str", # Optional. Detail view for + GetModelCatalogCard. + "type": "str" # Optional. Detail view for GetModelCatalogCard. } } # response body for status code(s): 404 @@ -252832,8 +266388,9 @@ def regenerate_model_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: cls: ClsType[JSON] = kwargs.pop("cls", None) - _request = build_genai_regenerate_model_api_key_request( - api_key_uuid=api_key_uuid, + _request = build_genai_get_model_catalog_card_request( + id=id, + model_id=model_id, headers=_headers, params=_params, ) @@ -252952,8 +266509,7 @@ def list_model_routers( ], "selection_policy": { "prefer": "str" # Optional. - One of: none, cheapest, fastest. Known values are: - "none", "cheapest", and "fastest". + One of: none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. @@ -253108,7 +266664,7 @@ def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -253138,8 +266694,7 @@ def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -253214,8 +266769,7 @@ def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -253282,7 +266836,7 @@ def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -253312,8 +266866,7 @@ def create_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -253496,8 +267049,7 @@ def list_model_router_presets( ], "selection_policy": { "prefer": "str" # Optional. - One of: none, cheapest, fastest. Known values are: - "none", "cheapest", and "fastest". + One of: none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. @@ -253657,7 +267209,7 @@ def list_model_router_task_presets( "name": "str", # Optional. Display name. "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "tags": [ "str" # Optional. Lightweight labels for filtering. @@ -253794,8 +267346,7 @@ def get_model_router(self, uuid: str, **kwargs: Any) -> JSON: ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -253934,7 +267485,7 @@ def update_model_router( body = { "description": "str", # Optional. Model router description. "fallback_models": [ - "str" # Optional. Fallback models. + {} # Optional. ], "name": "str", # Optional. Model router name. "policies": [ @@ -253949,7 +267500,7 @@ def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -253980,8 +267531,7 @@ def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -254059,8 +267609,7 @@ def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -254114,7 +267663,7 @@ def update_model_router( body = { "description": "str", # Optional. Model router description. "fallback_models": [ - "str" # Optional. Fallback models. + {} # Optional. ], "name": "str", # Optional. Model router name. "policies": [ @@ -254129,7 +267678,7 @@ def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: none, cheapest, - fastest. Known values are: "none", "cheapest", and "fastest". + fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -254160,8 +267709,7 @@ def update_model_router( ], "selection_policy": { "prefer": "str" # Optional. One of: - none, cheapest, fastest. Known values are: "none", - "cheapest", and "fastest". + none, cheapest, fastest. }, "task_slug": "str" # Optional. Task slug. } @@ -254793,8 +268341,27 @@ def list_openai_api_keys( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -254808,16 +268375,64 @@ def list_openai_api_keys( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -254998,211 +268613,358 @@ def create_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, - "created_at": "2020-02-20 00:00:00", # Optional. - Creation date / time. - "inference_name": "str", # Optional. Internally used - name. - "inference_version": "str", # Optional. Internally - used version. - "is_foundational": bool, # Optional. True if it is a - foundational model provided by do. - "kb_default_chunk_size": 0, # Optional. Default - chunking size limit to show in UI. - "kb_max_chunk_size": 0, # Optional. Maximum chunk - size limit of model. - "kb_min_chunk_size": 0, # Optional. Minimum chunking - size token limits if model supports KNOWLEDGEBASE usecase. - "metadata": {}, # Optional. Additional meta data. - "name": "str", # Optional. Name of the model. - "parent_uuid": "str", # Optional. Unique id of the - model, this model is based on. - "provider": "MODEL_PROVIDER_DIGITALOCEAN", # - Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known - values are: "MODEL_PROVIDER_DIGITALOCEAN", - "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". - "updated_at": "2020-02-20 00:00:00", # Optional. - Last modified. - "upload_complete": bool, # Optional. Model has been - fully uploaded. - "url": "str", # Optional. Download url. - "usecases": [ - "str" # Optional. Usecases of the model. + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], + "inference_name": "str", # Optional. Internally used + name. + "inference_version": "str", # Optional. Internally + used version. + "is_foundational": bool, # Optional. True if it is a + foundational model provided by do. + "kb_default_chunk_size": 0, # Optional. Default + chunking size limit to show in UI. + "kb_max_chunk_size": 0, # Optional. Maximum chunk + size limit of model. + "kb_min_chunk_size": 0, # Optional. Minimum chunking + size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). + "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, + "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. + "parent_uuid": "str", # Optional. Unique id of the + model, this model is based on. + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # + Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known + values are: "MODEL_PROVIDER_DIGITALOCEAN", + "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "upload_complete": bool, # Optional. Model has been + fully uploaded. + "url": "str", # Optional. Download url. + "usecases": [ + "str" # Optional. Usecases of the model. + ], + "uuid": "str", # Optional. Unique id. + "version": { + "major": 0, # Optional. Major version + number. + "minor": 0, # Optional. Minor version + number. + "patch": 0 # Optional. Patch version number. + } + } + ], + "name": "str", # Optional. Name. + "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated + date. + "uuid": "str" # Optional. Uuid. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @overload + def create_openai_api_key( + self, + body: Optional[IO[bytes]] = None, + *, + content_type: str = "application/json", + **kwargs: Any, + ) -> JSON: + # pylint: disable=line-too-long + """Create OpenAI API Key. + + To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. + + :param body: Default value is None. + :type body: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # response body for status code(s): 200 + response == { + "api_key_info": { + "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. + "created_by": "str", # Optional. Created by user id from DO. + "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. + "models": [ + { + "agreement": { + "description": "str", # Optional. Agreement + Description. + "name": "str", # Optional. Agreement + Description. + "url": "str", # Optional. Agreement + Description. + "uuid": "str" # Optional. Agreement + Description. + }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], + "inference_name": "str", # Optional. Internally used + name. + "inference_version": "str", # Optional. Internally + used version. + "is_foundational": bool, # Optional. True if it is a + foundational model provided by do. + "kb_default_chunk_size": 0, # Optional. Default + chunking size limit to show in UI. + "kb_max_chunk_size": 0, # Optional. Maximum chunk + size limit of model. + "kb_min_chunk_size": 0, # Optional. Minimum chunking + size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). + "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, + "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. + "parent_uuid": "str", # Optional. Unique id of the + model, this model is based on. + "provider": "MODEL_PROVIDER_DIGITALOCEAN", # + Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known + values are: "MODEL_PROVIDER_DIGITALOCEAN", + "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "upload_complete": bool, # Optional. Model has been + fully uploaded. + "url": "str", # Optional. Download url. + "usecases": [ + "str" # Optional. Usecases of the model. + ], + "uuid": "str", # Optional. Unique id. + "version": { + "major": 0, # Optional. Major version + number. + "minor": 0, # Optional. Minor version + number. + "patch": 0 # Optional. Patch version number. + } + } + ], + "name": "str", # Optional. Name. + "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated + date. + "uuid": "str" # Optional. Uuid. + } + } + # response body for status code(s): 404 + response == { + "id": "str", # A short identifier corresponding to the HTTP status code + returned. For example, the ID for a response returning a 404 status code would + be "not_found.". Required. + "message": "str", # A message providing additional information about the + error, including details to help resolve it when possible. Required. + "request_id": "str" # Optional. Optionally, some endpoints may include a + request ID that should be provided when reporting bugs or opening support + tickets to help identify the issue. + } + """ + + @distributed_trace + def create_openai_api_key( + self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any + ) -> JSON: + # pylint: disable=line-too-long + """Create OpenAI API Key. + + To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. + + :param body: Is either a JSON type or a IO[bytes] type. Default value is None. + :type body: JSON or IO[bytes] + :return: JSON object + :rtype: JSON + :raises ~azure.core.exceptions.HttpResponseError: + + Example: + .. code-block:: python + + # JSON input template you can fill out and use as your body input. + body = { + "api_key": "str", # Optional. OpenAI API key. + "name": "str" # Optional. Name of the key. + } + + # response body for status code(s): 200 + response == { + "api_key_info": { + "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. + "created_by": "str", # Optional. Created by user id from DO. + "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. + "models": [ + { + "agreement": { + "description": "str", # Optional. Agreement + Description. + "name": "str", # Optional. Agreement + Description. + "url": "str", # Optional. Agreement + Description. + "uuid": "str" # Optional. Agreement + Description. + }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). ], - "uuid": "str", # Optional. Unique id. - "version": { - "major": 0, # Optional. Major version - number. - "minor": 0, # Optional. Minor version - number. - "patch": 0 # Optional. Patch version number. - } - } - ], - "name": "str", # Optional. Name. - "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated - date. - "uuid": "str" # Optional. Uuid. - } - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @overload - def create_openai_api_key( - self, - body: Optional[IO[bytes]] = None, - *, - content_type: str = "application/json", - **kwargs: Any, - ) -> JSON: - # pylint: disable=line-too-long - """Create OpenAI API Key. - - To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. - - :param body: Default value is None. - :type body: IO[bytes] - :keyword content_type: Body Parameter content-type. Content type parameter for binary body. - Default value is "application/json". - :paramtype content_type: str - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "api_key_info": { - "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. - "created_by": "str", # Optional. Created by user id from DO. - "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. - "models": [ - { - "agreement": { - "description": "str", # Optional. Agreement - Description. - "name": "str", # Optional. Agreement - Description. - "url": "str", # Optional. Agreement - Description. - "uuid": "str" # Optional. Agreement - Description. - }, + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. - "inference_name": "str", # Optional. Internally used - name. - "inference_version": "str", # Optional. Internally - used version. - "is_foundational": bool, # Optional. True if it is a - foundational model provided by do. - "kb_default_chunk_size": 0, # Optional. Default - chunking size limit to show in UI. - "kb_max_chunk_size": 0, # Optional. Maximum chunk - size limit of model. - "kb_min_chunk_size": 0, # Optional. Minimum chunking - size token limits if model supports KNOWLEDGEBASE usecase. - "metadata": {}, # Optional. Additional meta data. - "name": "str", # Optional. Name of the model. - "parent_uuid": "str", # Optional. Unique id of the - model, this model is based on. - "provider": "MODEL_PROVIDER_DIGITALOCEAN", # - Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known - values are: "MODEL_PROVIDER_DIGITALOCEAN", - "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". - "updated_at": "2020-02-20 00:00:00", # Optional. - Last modified. - "upload_complete": bool, # Optional. Model has been - fully uploaded. - "url": "str", # Optional. Download url. - "usecases": [ - "str" # Optional. Usecases of the model. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } ], - "uuid": "str", # Optional. Unique id. - "version": { - "major": 0, # Optional. Major version - number. - "minor": 0, # Optional. Minor version - number. - "patch": 0 # Optional. Patch version number. - } - } - ], - "name": "str", # Optional. Name. - "updated_at": "2020-02-20 00:00:00", # Optional. Key last updated - date. - "uuid": "str" # Optional. Uuid. - } - } - # response body for status code(s): 404 - response == { - "id": "str", # A short identifier corresponding to the HTTP status code - returned. For example, the ID for a response returning a 404 status code would - be "not_found.". Required. - "message": "str", # A message providing additional information about the - error, including details to help resolve it when possible. Required. - "request_id": "str" # Optional. Optionally, some endpoints may include a - request ID that should be provided when reporting bugs or opening support - tickets to help identify the issue. - } - """ - - @distributed_trace - def create_openai_api_key( - self, body: Optional[Union[JSON, IO[bytes]]] = None, **kwargs: Any - ) -> JSON: - # pylint: disable=line-too-long - """Create OpenAI API Key. - - To create an OpenAI API key, send a POST request to ``/v2/gen-ai/openai/keys``. - - :param body: Is either a JSON type or a IO[bytes] type. Default value is None. - :type body: JSON or IO[bytes] - :return: JSON object - :rtype: JSON - :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "api_key": "str", # Optional. OpenAI API key. - "name": "str" # Optional. Name of the key. - } - - # response body for status code(s): 200 - response == { - "api_key_info": { - "created_at": "2020-02-20 00:00:00", # Optional. Key creation date. - "created_by": "str", # Optional. Created by user id from DO. - "deleted_at": "2020-02-20 00:00:00", # Optional. Key deleted date. - "models": [ - { - "agreement": { - "description": "str", # Optional. Agreement - Description. - "name": "str", # Optional. Agreement - Description. - "url": "str", # Optional. Agreement - Description. - "uuid": "str" # Optional. Agreement - Description. - }, - "created_at": "2020-02-20 00:00:00", # Optional. - Creation date / time. "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -255215,14 +268977,59 @@ def create_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -255388,8 +269195,27 @@ def get_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -255402,14 +269228,59 @@ def get_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -255577,8 +269448,27 @@ def update_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -255591,14 +269481,59 @@ def update_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -255682,8 +269617,27 @@ def update_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -255696,14 +269650,59 @@ def update_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -255789,8 +269788,27 @@ def update_openai_api_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -255803,14 +269821,59 @@ def update_openai_api_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -255977,8 +270040,27 @@ def delete_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -255991,14 +270073,59 @@ def delete_openai_api_key(self, api_key_uuid: str, **kwargs: Any) -> JSON: size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -256429,6 +270556,24 @@ def list_agents_by_openai_key( stream. }, "max_tokens": 0, # Optional. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -256440,8 +270585,27 @@ def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -256454,14 +270618,59 @@ def list_agents_by_openai_key( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -256499,8 +270708,27 @@ def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -256514,10 +270742,25 @@ def list_agents_by_openai_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -256525,6 +270768,42 @@ def list_agents_by_openai_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -256555,6 +270834,45 @@ def list_agents_by_openai_key( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -256575,8 +270893,27 @@ def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -256590,10 +270927,25 @@ def list_agents_by_openai_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -256601,6 +270953,42 @@ def list_agents_by_openai_key( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -256634,6 +271022,8 @@ def list_agents_by_openai_key( "project_id": "str", # Optional. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -256814,8 +271204,27 @@ def list_agents_by_openai_key( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -256829,16 +271238,64 @@ def list_agents_by_openai_key( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -256882,6 +271339,8 @@ def list_agents_by_openai_key( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -256924,6 +271383,13 @@ def list_agents_by_openai_key( 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -256954,6 +271420,16 @@ def list_agents_by_openai_key( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -256967,8 +271443,10 @@ def list_agents_by_openai_key( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -258169,6 +272647,27 @@ def list_workspaces(self, **kwargs: Any) -> JSON: Name of the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. + Optional list of allowed tool names to expose from + this server. + ], + "authorization": "str", # + Optional. Optional authorization header value for the MCP + server. + "headers": { + "str": "str" # + Optional. Optional additional headers to send to the + MCP server. + }, + "server_label": "str", # + Optional. A label identifying this MCP server. + "server_url": "str" # + Optional. The URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # @@ -258180,8 +272679,27 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -258195,10 +272713,25 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -258206,6 +272739,42 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -258248,9 +272817,30 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": + {}, # Optional. Benchmark scores for this model, + stored as arbitrary JSON. + "capabilities": [ + "str" # + Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": + "str", # Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities supported + by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": @@ -258265,10 +272855,26 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": + "str", # Optional. Lifecycle status of the model + (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" + # Optional. Input/output modalities. + ], + "output": [ + "str" + # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": + 0.0, # Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. @@ -258278,6 +272884,43 @@ def list_workspaces(self, **kwargs: Any) -> JSON: are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": + [ + "str" # + Optional. Available reasoning efforts for this + model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended + thinking (Anthropic models). + "type": "str", # + Optional. Model type (chat, embedding, image, + reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": @@ -258310,6 +272953,46 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" # Optional. Task name. + }, + "models": [ + "str" + # Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # Optional. One of: none, + cheapest, fastest. + }, + "task_slug": + "str" # Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", + # Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of + the model router. + "regions": [ + "str" # Optional. Target + regions for the router. + ], + "updated_at": "2020-02-20 00:00:00", + # Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", @@ -258330,9 +273013,30 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": + {}, # Optional. Benchmark scores for this model, + stored as arbitrary JSON. + "capabilities": [ + "str" # + Optional. High-level capabilities (e.g. + tool_calling, vision, streaming). + ], + "context_window": + "str", # Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. Capabilities supported + by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The endpoint + path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": @@ -258347,10 +273051,26 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": + "str", # Optional. Lifecycle status of the model + (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" + # Optional. Input/output modalities. + ], + "output": [ + "str" + # Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": + 0.0, # Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. @@ -258360,6 +273080,43 @@ def list_workspaces(self, **kwargs: Any) -> JSON: are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": + [ + "str" # + Optional. Available reasoning efforts for this + model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, # Optional. Maximum allowed + value (for type="number"). + "min": 0.0, # Optional. Minimum allowed + value (for type="number"). + "name": "str", # Optional. Setting key name + (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. Allowed values for + dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. Step increment for + numeric settings (for type="number"). + "type": "str" # Optional. Setting value + type: "number" or "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended + thinking (Anthropic models). + "type": "str", # + Optional. Model type (chat, embedding, image, + reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": @@ -258393,6 +273150,8 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The + reasoning effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is @@ -258593,8 +273352,31 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -258609,10 +273391,25 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -258620,6 +273417,44 @@ def list_workspaces(self, **kwargs: Any) -> JSON: is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -258666,6 +273501,9 @@ def list_workspaces(self, **kwargs: Any) -> JSON: # Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The + thinking token budget for Anthropic extended thinking (0 = + disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -258708,6 +273546,13 @@ def list_workspaces(self, **kwargs: Any) -> JSON: # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value + is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The @@ -258738,6 +273583,15 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent evaluation + or model evaluation. For backwards compatibility, + UNSPECIFIED defaults to agent metrics only in list + operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", + and "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -258751,7 +273605,8 @@ def list_workspaces(self, **kwargs: Any) -> JSON: "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", - and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", + and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -259227,6 +274082,24 @@ def create_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -259238,8 +274111,27 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -259253,16 +274145,64 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -259303,8 +274243,31 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -259319,10 +274282,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -259330,6 +274308,44 @@ def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -259360,6 +274376,48 @@ def create_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -259380,8 +274438,31 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -259396,10 +274477,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -259407,6 +274503,44 @@ def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -259440,6 +274574,8 @@ def create_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -259631,8 +274767,27 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -259646,10 +274801,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -259657,6 +274827,42 @@ def create_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -259703,6 +274909,8 @@ def create_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -259742,6 +274950,12 @@ def create_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -259768,6 +274982,14 @@ def create_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -259780,8 +275002,9 @@ def create_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -260167,6 +275390,24 @@ def create_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -260178,8 +275419,27 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -260193,16 +275453,64 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -260243,8 +275551,31 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -260259,10 +275590,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -260270,6 +275616,44 @@ def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -260300,6 +275684,48 @@ def create_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -260320,8 +275746,31 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -260336,10 +275785,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -260347,6 +275811,44 @@ def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -260380,6 +275882,8 @@ def create_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -260571,8 +276075,27 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -260586,10 +276109,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -260597,6 +276135,42 @@ def create_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -260643,6 +276217,8 @@ def create_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -260682,6 +276258,12 @@ def create_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -260708,6 +276290,14 @@ def create_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -260720,8 +276310,9 @@ def create_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -261109,6 +276700,24 @@ def create_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -261120,8 +276729,27 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -261135,16 +276763,64 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -261185,8 +276861,31 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -261201,10 +276900,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -261212,6 +276926,44 @@ def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -261242,6 +276994,48 @@ def create_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -261262,8 +277056,31 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -261278,10 +277095,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -261289,6 +277121,44 @@ def create_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -261322,6 +277192,8 @@ def create_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -261513,8 +277385,27 @@ def create_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -261528,10 +277419,25 @@ def create_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -261539,6 +277445,42 @@ def create_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -261585,6 +277527,8 @@ def create_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -261624,6 +277568,12 @@ def create_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -261650,6 +277600,14 @@ def create_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -261662,8 +277620,9 @@ def create_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -262133,6 +278092,24 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -262144,8 +278121,27 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -262159,16 +278155,64 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -262209,8 +278253,31 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -262225,10 +278292,25 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -262236,6 +278318,44 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -262266,6 +278386,48 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -262286,8 +278448,31 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -262302,10 +278487,25 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -262313,6 +278513,44 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -262346,6 +278584,8 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -262537,8 +278777,27 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -262552,10 +278811,25 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -262563,6 +278837,42 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -262609,6 +278919,8 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -262648,6 +278960,12 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -262674,6 +278992,14 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -262686,8 +279012,9 @@ def get_workspace(self, workspace_uuid: str, **kwargs: Any) -> JSON: "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -263160,6 +279487,24 @@ def update_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -263171,8 +279516,27 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -263186,16 +279550,64 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -263236,8 +279648,31 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -263252,10 +279687,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -263263,6 +279713,44 @@ def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -263293,6 +279781,48 @@ def update_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -263313,8 +279843,31 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -263329,10 +279882,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -263340,6 +279908,44 @@ def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -263373,6 +279979,8 @@ def update_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -263564,8 +280172,27 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -263579,10 +280206,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -263590,6 +280232,42 @@ def update_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -263636,6 +280314,8 @@ def update_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -263675,6 +280355,12 @@ def update_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -263701,6 +280387,14 @@ def update_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -263713,8 +280407,9 @@ def update_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -264103,6 +280798,24 @@ def update_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -264114,8 +280827,27 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -264129,16 +280861,64 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -264179,8 +280959,31 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -264195,10 +280998,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -264206,6 +281024,44 @@ def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -264236,6 +281092,48 @@ def update_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -264256,8 +281154,31 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -264272,10 +281193,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -264283,6 +281219,44 @@ def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -264316,6 +281290,8 @@ def update_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -264507,8 +281483,27 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -264522,10 +281517,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -264533,6 +281543,42 @@ def update_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -264579,6 +281625,8 @@ def update_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -264618,6 +281666,12 @@ def update_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -264644,6 +281698,14 @@ def update_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -264656,8 +281718,9 @@ def update_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -265048,6 +282111,24 @@ def update_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -265059,8 +282140,27 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -265074,16 +282174,64 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -265124,8 +282272,31 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -265140,10 +282311,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -265151,6 +282337,44 @@ def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -265181,6 +282405,48 @@ def update_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -265201,8 +282467,31 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -265217,10 +282506,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -265228,6 +282532,44 @@ def update_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -265261,6 +282603,8 @@ def update_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -265452,8 +282796,27 @@ def update_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -265467,10 +282830,25 @@ def update_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -265478,6 +282856,42 @@ def update_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -265524,6 +282938,8 @@ def update_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -265563,6 +282979,12 @@ def update_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -265589,6 +283011,14 @@ def update_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -265601,8 +283031,9 @@ def update_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -266182,6 +283613,24 @@ def list_agents_by_workspace( stream. }, "max_tokens": 0, # Optional. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional list of + allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. Optional + authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. Optional + additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A label + identifying this MCP server. + "server_url": "str" # Optional. The URL of + the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. Agreement @@ -266193,8 +283642,27 @@ def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark scores + for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level capabilities + (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context window + size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. input_text, + output_text, input_image). + ], + "endpoint": "str" # Optional. The + endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. Internally @@ -266207,14 +283675,59 @@ def list_agents_by_workspace( size limit of model. "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. Lifecycle + status of the model (internal, public-preview, active, deprecated, + end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. Input/output + modalities. + ], + "output": [ + "str" # Optional. Input/output + modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. Parameter count + in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available reasoning + efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. + String default value (for type="dropdown", e.g. "medium"). + "default_value": 0.0, # Optional. + Numeric default value (for type="number"). + "max": 0.0, # Optional. Maximum + allowed value (for type="number"). + "min": 0.0, # Optional. Minimum + allowed value (for type="number"). + "name": "str", # Optional. Setting + key name (e.g. "max_tokens", "temperature", "resolution"). + "options": [ + "str" # Optional. Allowed + values for dropdown selections (for type="dropdown"). + ], + "step": 0.0, # Optional. Step + increment for numeric settings (for type="number"). + "type": "str" # Optional. Setting + value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this model + supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model has been @@ -266252,8 +283765,27 @@ def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -266267,10 +283799,25 @@ def list_agents_by_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -266278,6 +283825,42 @@ def list_agents_by_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -266308,6 +283891,45 @@ def list_agents_by_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. Router-level + fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", + # Optional. Short task description. + "name": "str" # + Optional. Task name. + }, + "models": [ + "str" # Optional. + Models assigned to the task. + ], + "selection_policy": { + "prefer": "str" # + Optional. One of: none, cheapest, fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # Optional. + Creation date / time. + "description": "str", # Optional. Description. + "name": "str", # Optional. Name of the model router. + "regions": [ + "str" # Optional. Target regions for the + router. + ], + "updated_at": "2020-02-20 00:00:00", # Optional. + Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # Optional. Key @@ -266328,8 +283950,27 @@ def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -266343,10 +283984,25 @@ def list_agents_by_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -266354,6 +284010,42 @@ def list_agents_by_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -266387,6 +284079,8 @@ def list_agents_by_workspace( "project_id": "str", # Optional. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning effort + for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * RETRIEVAL_METHOD_UNKNOWN: @@ -266567,8 +284261,27 @@ def list_agents_by_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -266582,16 +284295,64 @@ def list_agents_by_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -266635,6 +284396,8 @@ def list_agents_by_workspace( agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking token + budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -266677,6 +284440,13 @@ def list_agents_by_workspace( 00:00:00", # Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default + value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known values + are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # @@ -266707,6 +284477,16 @@ def list_agents_by_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default + value is "EVALUATION_SCOPE_UNSPECIFIED". Scope that + determines whether a metric belongs to agent + evaluation or model evaluation. For backwards + compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", + "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -266720,8 +284500,10 @@ def list_agents_by_workspace( "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -267221,6 +285003,24 @@ def update_agents_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -267232,8 +285032,27 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -267247,16 +285066,64 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -267297,8 +285164,31 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -267313,10 +285203,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -267324,6 +285229,44 @@ def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -267354,6 +285297,48 @@ def update_agents_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -267374,8 +285359,31 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -267390,10 +285398,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -267401,6 +285424,44 @@ def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -267434,6 +285495,8 @@ def update_agents_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -267625,8 +285688,27 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -267640,10 +285722,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -267651,6 +285748,42 @@ def update_agents_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -267697,6 +285830,8 @@ def update_agents_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -267736,6 +285871,12 @@ def update_agents_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -267762,6 +285903,14 @@ def update_agents_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -267774,8 +285923,9 @@ def update_agents_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -268164,6 +286314,24 @@ def update_agents_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -268175,8 +286343,27 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -268190,16 +286377,64 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -268240,8 +286475,31 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -268256,10 +286514,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -268267,6 +286540,44 @@ def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -268297,6 +286608,48 @@ def update_agents_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -268317,8 +286670,31 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -268333,10 +286709,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -268344,6 +286735,44 @@ def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -268377,6 +286806,8 @@ def update_agents_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -268568,8 +286999,27 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -268583,10 +287033,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -268594,6 +287059,42 @@ def update_agents_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -268640,6 +287141,8 @@ def update_agents_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -268679,6 +287182,12 @@ def update_agents_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -268705,6 +287214,14 @@ def update_agents_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -268717,8 +287234,9 @@ def update_agents_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -269110,6 +287628,24 @@ def update_agents_workspace( the log stream. }, "max_tokens": 0, # Optional. Agents. + "mcp_servers": [ + { + "allowed_tools": [ + "str" # Optional. Optional + list of allowed tool names to expose from this server. + ], + "authorization": "str", # Optional. + Optional authorization header value for the MCP server. + "headers": { + "str": "str" # Optional. + Optional additional headers to send to the MCP server. + }, + "server_label": "str", # Optional. A + label identifying this MCP server. + "server_url": "str" # Optional. The + URL of the MCP server. + } + ], "model": { "agreement": { "description": "str", # Optional. @@ -269121,8 +287657,27 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. Benchmark + scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. Context + window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # Optional. + Capabilities supported by this endpoint (e.g. + input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # Optional. @@ -269136,16 +287691,64 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # Optional. + Lifecycle status of the model (internal, public-preview, active, + deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": "MODEL_PROVIDER_DIGITALOCEAN", # Optional. Default value is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # + Optional. String default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # + Optional. Numeric default value (for type="number"). + "max": 0.0, # Optional. + Maximum allowed value (for type="number"). + "min": 0.0, # Optional. + Minimum allowed value (for type="number"). + "name": "str", # Optional. + Setting key name (e.g. "max_tokens", "temperature", + "resolution"). + "options": [ + "str" # Optional. + Allowed values for dropdown selections (for + type="dropdown"). + ], + "step": 0.0, # Optional. + Step increment for numeric settings (for type="number"). + "type": "str" # Optional. + Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. Whether this + model supports extended thinking (Anthropic models). + "type": "str", # Optional. Model type (chat, + embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. Model @@ -269186,8 +287789,31 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -269202,10 +287828,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -269213,6 +287854,44 @@ def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -269243,6 +287922,48 @@ def update_agents_workspace( "updated_at": "2020-02-20 00:00:00" # Optional. Key last updated date. }, + "model_router": { + "config": { + "fallback_models": [ + "str" # Optional. + Router-level fallback models. + ], + "policies": [ + { + "custom_task": { + "description": "str", # Optional. Short task + description. + "name": "str" + # Optional. Task name. + }, + "models": [ + "str" # + Optional. Models assigned to the task. + ], + "selection_policy": { + "prefer": + "str" # Optional. One of: none, cheapest, + fastest. + }, + "task_slug": "str" # + Optional. Task slug. + } + ] + }, + "created_at": "2020-02-20 00:00:00", # + Optional. Creation date / time. + "description": "str", # Optional. + Description. + "name": "str", # Optional. Name of the model + router. + "regions": [ + "str" # Optional. Target regions for + the router. + ], + "updated_at": "2020-02-20 00:00:00", # + Optional. Last modified. + "uuid": "str" # Optional. Unique id. + }, "name": "str", # Optional. Agent name. "openai_api_key": { "created_at": "2020-02-20 00:00:00", # @@ -269263,8 +287984,31 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # + Optional. Benchmark scores for this model, stored as + arbitrary JSON. + "capabilities": [ + "str" # Optional. + High-level capabilities (e.g. tool_calling, vision, + streaming). + ], + "context_window": "str", # + Optional. Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" + # Optional. Capabilities supported by this + endpoint (e.g. input_text, output_text, + input_image). + ], + "endpoint": + "str" # Optional. The endpoint path (e.g. + /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", @@ -269279,10 +288023,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # + Optional. Input/output modalities. + ], + "output": [ + "str" # + Optional. Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # + Optional. Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -269290,6 +288049,44 @@ def update_agents_workspace( is "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. + Available reasoning efforts for this model. + ], + "settings": [ + { + "default_string": "str", # Optional. String + default value (for type="dropdown", e.g. + "medium"). + "default_value": 0.0, # Optional. Numeric + default value (for type="number"). + "max": 0.0, + # Optional. Maximum allowed value (for + type="number"). + "min": 0.0, + # Optional. Minimum allowed value (for + type="number"). + "name": + "str", # Optional. Setting key name (e.g. + "max_tokens", "temperature", "resolution"). + "options": [ + "str" + # Optional. Allowed values for dropdown + selections (for type="dropdown"). + ], + "step": 0.0, + # Optional. Step increment for numeric settings + (for type="number"). + "type": "str" + # Optional. Setting value type: "number" or + "dropdown". + } + ], + "thinking": bool, # + Optional. Whether this model supports extended thinking + (Anthropic models). + "type": "str", # Optional. + Model type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # @@ -269323,6 +288120,8 @@ def update_agents_workspace( "project_id": "str", # Optional. Agents. "provide_citations": bool, # Optional. Whether the agent should provide in-response citations. + "reasoning_effort": "str", # Optional. The reasoning + effort for the agent. "region": "str", # Optional. Region code. "retrieval_method": "RETRIEVAL_METHOD_UNKNOWN", # Optional. Default value is "RETRIEVAL_METHOD_UNKNOWN". * @@ -269514,8 +288313,27 @@ def update_agents_workspace( "uuid": "str" # Optional. Agreement Description. }, + "benchmark_score": {}, # Optional. + Benchmark scores for this model, stored as arbitrary JSON. + "capabilities": [ + "str" # Optional. High-level + capabilities (e.g. tool_calling, vision, streaming). + ], + "context_window": "str", # Optional. + Context window size in tokens. "created_at": "2020-02-20 00:00:00", # Optional. Creation date / time. + "endpoints": [ + { + "capabilities": [ + "str" # + Optional. Capabilities supported by this endpoint + (e.g. input_text, output_text, input_image). + ], + "endpoint": "str" # + Optional. The endpoint path (e.g. /chat/responses). + } + ], "inference_name": "str", # Optional. Internally used name. "inference_version": "str", # @@ -269529,10 +288347,25 @@ def update_agents_workspace( "kb_min_chunk_size": 0, # Optional. Minimum chunking size token limits if model supports KNOWLEDGEBASE usecase. + "lifecycle_status": "str", # + Optional. Lifecycle status of the model (internal, + public-preview, active, deprecated, end_of_life). "metadata": {}, # Optional. Additional meta data. + "modalities": { + "input": [ + "str" # Optional. + Input/output modalities. + ], + "output": [ + "str" # Optional. + Input/output modalities. + ] + }, "name": "str", # Optional. Name of the model. + "parameter_count": 0.0, # Optional. + Parameter count in billions. "parent_uuid": "str", # Optional. Unique id of the model, this model is based on. "provider": @@ -269540,6 +288373,42 @@ def update_agents_workspace( "MODEL_PROVIDER_DIGITALOCEAN". Known values are: "MODEL_PROVIDER_DIGITALOCEAN", "MODEL_PROVIDER_ANTHROPIC", and "MODEL_PROVIDER_OPENAI". + "reasoning_efforts": [ + "str" # Optional. Available + reasoning efforts for this model. + ], + "settings": [ + { + "default_string": + "str", # Optional. String default value (for + type="dropdown", e.g. "medium"). + "default_value": 0.0, + # Optional. Numeric default value (for + type="number"). + "max": 0.0, # + Optional. Maximum allowed value (for type="number"). + "min": 0.0, # + Optional. Minimum allowed value (for type="number"). + "name": "str", # + Optional. Setting key name (e.g. "max_tokens", + "temperature", "resolution"). + "options": [ + "str" # + Optional. Allowed values for dropdown selections + (for type="dropdown"). + ], + "step": 0.0, # + Optional. Step increment for numeric settings (for + type="number"). + "type": "str" # + Optional. Setting value type: "number" or "dropdown". + } + ], + "thinking": bool, # Optional. + Whether this model supports extended thinking (Anthropic + models). + "type": "str", # Optional. Model + type (chat, embedding, image, reasoning, coding). "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. "upload_complete": bool, # Optional. @@ -269586,6 +288455,8 @@ def update_agents_workspace( Optional. The agent template's last updated date. "uuid": "str" # Optional. Unique id. }, + "thinking_token_budget": 0, # Optional. The thinking + token budget for Anthropic extended thinking (0 = disabled). "top_p": 0.0, # Optional. Agents. "updated_at": "2020-02-20 00:00:00", # Optional. Last modified. @@ -269625,6 +288496,12 @@ def update_agents_workspace( Optional. Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": + "EVALUATION_DATASET_TYPE_UNKNOWN", # Optional. Default value is + "EVALUATION_DATASET_TYPE_UNKNOWN". Known values are: + "EVALUATION_DATASET_TYPE_UNKNOWN", "EVALUATION_DATASET_TYPE_ADK", + "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of @@ -269651,6 +288528,14 @@ def update_agents_workspace( "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. Evaluations. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether + a metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent + metrics only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. @@ -269663,8 +288548,9 @@ def update_agents_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: - "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", and - "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_UNSPECIFIED", "METRIC_TYPE_GENERAL_QUALITY", + "METRIC_TYPE_RAG_AND_TOOL", "METRIC_TYPE_MODEL_QUALITY", and + "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. Evaluations. "metric_value_type": @@ -269846,6 +288732,11 @@ def list_evaluation_test_cases_by_workspace( Time created at. "dataset_name": "str", # Optional. Name of the dataset. + "dataset_type": "EVALUATION_DATASET_TYPE_UNKNOWN", # + Optional. Default value is "EVALUATION_DATASET_TYPE_UNKNOWN". Known + values are: "EVALUATION_DATASET_TYPE_UNKNOWN", + "EVALUATION_DATASET_TYPE_ADK", "EVALUATION_DATASET_TYPE_NON_ADK", and + "EVALUATION_DATASET_TYPE_MODEL". "dataset_uuid": "str", # Optional. UUID of the dataset. "file_size": "str", # Optional. The size of the @@ -269869,6 +288760,14 @@ def list_evaluation_test_cases_by_workspace( "METRIC_CATEGORY_CONTEXT_QUALITY", and "METRIC_CATEGORY_MODEL_FIT". "description": "str", # Optional. + "evaluation_scope": + "EVALUATION_SCOPE_UNSPECIFIED", # Optional. Default value is + "EVALUATION_SCOPE_UNSPECIFIED". Scope that determines whether a + metric belongs to agent evaluation or model evaluation. For + backwards compatibility, UNSPECIFIED defaults to agent metrics + only in list operations. Known values are: + "EVALUATION_SCOPE_UNSPECIFIED", "EVALUATION_SCOPE_AGENT", and + "EVALUATION_SCOPE_MODEL". "inverted": bool, # Optional. If true, the metric is inverted, meaning that a lower value is better. "is_metric_goal": bool, # Optional. @@ -269877,7 +288776,8 @@ def list_evaluation_test_cases_by_workspace( "metric_type": "METRIC_TYPE_UNSPECIFIED", # Optional. Default value is "METRIC_TYPE_UNSPECIFIED". Known values are: "METRIC_TYPE_UNSPECIFIED", - "METRIC_TYPE_GENERAL_QUALITY", and "METRIC_TYPE_RAG_AND_TOOL". + "METRIC_TYPE_GENERAL_QUALITY", "METRIC_TYPE_RAG_AND_TOOL", + "METRIC_TYPE_MODEL_QUALITY", and "METRIC_TYPE_MODEL_SAFETY". "metric_uuid": "str", # Optional. "metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED", # Optional. Default value is