diff --git a/vertexai/_genai/a2a_task_events.py b/vertexai/_genai/a2a_task_events.py index 8113427370..15e77b98cb 100644 --- a/vertexai/_genai/a2a_task_events.py +++ b/vertexai/_genai/a2a_task_events.py @@ -167,7 +167,24 @@ def append( ) return_value = types.AppendAgentEngineTaskEventResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -232,7 +249,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineTaskEventsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -334,7 +368,24 @@ async def append( ) return_value = types.AppendAgentEngineTaskEventResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -401,7 +452,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineTaskEventsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/a2a_tasks.py b/vertexai/_genai/a2a_tasks.py index 5a83a16aac..c0a955dc05 100644 --- a/vertexai/_genai/a2a_tasks.py +++ b/vertexai/_genai/a2a_tasks.py @@ -187,7 +187,24 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.A2aTask._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -252,7 +269,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineTasksResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -321,7 +355,24 @@ def create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.A2aTask._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -433,7 +484,24 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.A2aTask._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -500,7 +568,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineTasksResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -571,7 +656,24 @@ async def create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.A2aTask._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/agent_engines.py b/vertexai/_genai/agent_engines.py index cf6e96c943..4272e5b34e 100644 --- a/vertexai/_genai/agent_engines.py +++ b/vertexai/_genai/agent_engines.py @@ -448,7 +448,24 @@ def _check_query_job( response_dict = _CheckQueryJobResult_from_vertex(response_dict) return_value = types.CheckQueryJobResult._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -503,7 +520,24 @@ def _run_query_job( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -554,7 +588,24 @@ def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -624,7 +675,24 @@ def _delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -674,7 +742,24 @@ def _get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ReasoningEngine._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -723,7 +808,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListReasoningEnginesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -772,7 +874,24 @@ def _get_agent_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -822,7 +941,24 @@ def _query( response_dict = {} if not response.body else json.loads(response.body) return_value = types.QueryReasoningEngineResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -874,7 +1010,24 @@ def _update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2517,7 +2670,24 @@ async def _check_query_job( response_dict = _CheckQueryJobResult_from_vertex(response_dict) return_value = types.CheckQueryJobResult._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2574,7 +2744,24 @@ async def _run_query_job( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2627,7 +2814,24 @@ async def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2699,7 +2903,24 @@ async def _delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2751,7 +2972,24 @@ async def _get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ReasoningEngine._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2802,7 +3040,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListReasoningEnginesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2853,7 +3108,24 @@ async def _get_agent_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2905,7 +3177,24 @@ async def _query( response_dict = {} if not response.body else json.loads(response.body) return_value = types.QueryReasoningEngineResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2959,7 +3248,24 @@ async def _update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/datasets.py b/vertexai/_genai/datasets.py index 462ec4566c..f5ef7793b1 100644 --- a/vertexai/_genai/datasets.py +++ b/vertexai/_genai/datasets.py @@ -299,7 +299,24 @@ def _assemble_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -372,7 +389,24 @@ def _assess_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -435,7 +469,24 @@ def _create_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -487,7 +538,24 @@ def _delete_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -540,7 +608,24 @@ def _get_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -599,7 +684,24 @@ def _get_multimodal_dataset_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -650,7 +752,24 @@ def _list_multimodal_datasets( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListMultimodalDatasetsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -711,7 +830,24 @@ def _update_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1366,7 +1502,24 @@ async def _assemble_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1441,7 +1594,24 @@ async def _assess_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1506,7 +1676,24 @@ async def _create_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1560,7 +1747,24 @@ async def _delete_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1615,7 +1819,24 @@ async def _get_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1676,7 +1897,24 @@ async def _get_multimodal_dataset_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1729,7 +1967,24 @@ async def _list_multimodal_datasets( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListMultimodalDatasetsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1792,7 +2047,24 @@ async def _update_multimodal_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MultimodalDataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/evals.py b/vertexai/_genai/evals.py index 0adf80cd6e..ece5486c35 100644 --- a/vertexai/_genai/evals.py +++ b/vertexai/_genai/evals.py @@ -963,7 +963,24 @@ def _create_evaluation_item( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationItem._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1023,7 +1040,24 @@ def _create_evaluation_metric( response_dict = _EvaluationMetric_from_vertex(response_dict) return_value = types.EvaluationMetric._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1091,7 +1125,24 @@ def _create_evaluation_run( response_dict = _EvaluationRun_from_vertex(response_dict) return_value = types.EvaluationRun._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1146,7 +1197,24 @@ def _create_evaluation_set( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationSet._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1231,7 +1299,24 @@ def _evaluate_instances( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluateInstancesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1292,7 +1377,24 @@ def _generate_user_scenarios( response_dict = {} if not response.body else json.loads(response.body) return_value = types.GenerateUserScenariosResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1351,7 +1453,24 @@ def _generate_loss_clusters( response_dict = {} if not response.body else json.loads(response.body) return_value = types.GenerateLossClustersOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1412,7 +1531,24 @@ def _generate_rubrics( response_dict = {} if not response.body else json.loads(response.body) return_value = types.GenerateInstanceRubricsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1468,7 +1604,24 @@ def _get_evaluation_metric( response_dict = _EvaluationMetric_from_vertex(response_dict) return_value = types.EvaluationMetric._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1521,7 +1674,24 @@ def _get_evaluation_run( response_dict = _EvaluationRun_from_vertex(response_dict) return_value = types.EvaluationRun._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1571,7 +1741,24 @@ def _get_evaluation_set( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationSet._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1621,7 +1808,24 @@ def _get_evaluation_item( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationItem._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1673,7 +1877,24 @@ def _list_evaluation_metrics( response_dict = _ListEvaluationMetricsResponse_from_vertex(response_dict) return_value = types.ListEvaluationMetricsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2577,7 +2798,24 @@ async def _create_evaluation_item( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationItem._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2639,7 +2877,24 @@ async def _create_evaluation_metric( response_dict = _EvaluationMetric_from_vertex(response_dict) return_value = types.EvaluationMetric._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2709,7 +2964,24 @@ async def _create_evaluation_run( response_dict = _EvaluationRun_from_vertex(response_dict) return_value = types.EvaluationRun._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2766,7 +3038,24 @@ async def _create_evaluation_set( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationSet._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2853,7 +3142,24 @@ async def _evaluate_instances( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluateInstancesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2916,7 +3222,24 @@ async def _generate_user_scenarios( response_dict = {} if not response.body else json.loads(response.body) return_value = types.GenerateUserScenariosResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2977,7 +3300,24 @@ async def _generate_loss_clusters( response_dict = {} if not response.body else json.loads(response.body) return_value = types.GenerateLossClustersOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -3040,7 +3380,24 @@ async def _generate_rubrics( response_dict = {} if not response.body else json.loads(response.body) return_value = types.GenerateInstanceRubricsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -3098,7 +3455,24 @@ async def _get_evaluation_metric( response_dict = _EvaluationMetric_from_vertex(response_dict) return_value = types.EvaluationMetric._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -3153,7 +3527,24 @@ async def _get_evaluation_run( response_dict = _EvaluationRun_from_vertex(response_dict) return_value = types.EvaluationRun._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -3205,7 +3596,24 @@ async def _get_evaluation_set( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationSet._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -3257,7 +3665,24 @@ async def _get_evaluation_item( response_dict = {} if not response.body else json.loads(response.body) return_value = types.EvaluationItem._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -3311,7 +3736,24 @@ async def _list_evaluation_metrics( response_dict = _ListEvaluationMetricsResponse_from_vertex(response_dict) return_value = types.ListEvaluationMetricsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/memories.py b/vertexai/_genai/memories.py index 2d4b1b180c..23aa9c0016 100644 --- a/vertexai/_genai/memories.py +++ b/vertexai/_genai/memories.py @@ -514,7 +514,24 @@ def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -577,7 +594,24 @@ def delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -646,7 +680,24 @@ def _generate( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineGenerateMemoriesOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -707,7 +758,24 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Memory._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -762,7 +830,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListReasoningEnginesMemoriesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -813,7 +898,24 @@ def _get_memory_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -864,7 +966,24 @@ def _get_generate_memories_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineGenerateMemoriesOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -929,7 +1048,24 @@ def _retrieve( response_dict = {} if not response.body else json.loads(response.body) return_value = types.RetrieveMemoriesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -986,7 +1122,24 @@ def _rollback( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineRollbackMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1045,7 +1198,24 @@ def _update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1106,7 +1276,24 @@ def _purge( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEnginePurgeMemoriesOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1474,7 +1661,24 @@ async def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1539,7 +1743,24 @@ async def delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1610,7 +1831,24 @@ async def _generate( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineGenerateMemoriesOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1673,7 +1911,24 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Memory._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1730,7 +1985,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListReasoningEnginesMemoriesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1783,7 +2055,24 @@ async def _get_memory_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1836,7 +2125,24 @@ async def _get_generate_memories_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineGenerateMemoriesOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1903,7 +2209,24 @@ async def _retrieve( response_dict = {} if not response.body else json.loads(response.body) return_value = types.RetrieveMemoriesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1962,7 +2285,24 @@ async def _rollback( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineRollbackMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2023,7 +2363,24 @@ async def _update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineMemoryOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2086,7 +2443,24 @@ async def _purge( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEnginePurgeMemoriesOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/memory_revisions.py b/vertexai/_genai/memory_revisions.py index c71c5cfcc7..07b20ed979 100644 --- a/vertexai/_genai/memory_revisions.py +++ b/vertexai/_genai/memory_revisions.py @@ -141,7 +141,24 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MemoryRevision._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -206,7 +223,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineMemoryRevisionsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -301,7 +335,24 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.MemoryRevision._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -368,7 +419,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineMemoryRevisionsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/prompt_optimizer.py b/vertexai/_genai/prompt_optimizer.py index ec62e6528c..5c3aa1aba6 100644 --- a/vertexai/_genai/prompt_optimizer.py +++ b/vertexai/_genai/prompt_optimizer.py @@ -252,7 +252,24 @@ def _optimize_prompt( response_dict = {} if not response.body else json.loads(response.body) return_value = types.OptimizeResponseEndpoint._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -308,7 +325,24 @@ def _create_custom_job_resource( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -361,7 +395,24 @@ def _get_custom_job( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -607,7 +658,24 @@ async def _optimize_prompt( response_dict = {} if not response.body else json.loads(response.body) return_value = types.OptimizeResponseEndpoint._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -665,7 +733,24 @@ async def _create_custom_job_resource( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -720,7 +805,24 @@ async def _get_custom_job( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/prompts.py b/vertexai/_genai/prompts.py index 6ed6da8bca..02fdf463d9 100644 --- a/vertexai/_genai/prompts.py +++ b/vertexai/_genai/prompts.py @@ -513,7 +513,24 @@ def _create_dataset_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -574,7 +591,24 @@ def _create_dataset_version_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -627,7 +661,24 @@ def _get_dataset_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Dataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -684,7 +735,24 @@ def _get_dataset_version_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetVersion._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -741,7 +809,24 @@ def _get_dataset_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -790,7 +875,24 @@ def _list_prompts( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListDatasetsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -849,7 +951,24 @@ def _list_versions( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListDatasetVersionsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -895,7 +1014,24 @@ def _delete_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeletePromptOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -950,7 +1086,24 @@ def _delete_dataset_version( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeletePromptVersionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1007,7 +1160,24 @@ def _restore_version( response_dict = {} if not response.body else json.loads(response.body) return_value = types.RestoreVersionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1072,7 +1242,24 @@ def _update_dataset_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Dataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1128,7 +1315,24 @@ def _create_custom_job_resource( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1181,7 +1385,24 @@ def _get_custom_job( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1234,7 +1455,24 @@ def _optimize( response_dict = {} if not response.body else json.loads(response.body) return_value = types.OptimizeResponseEndpoint._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2107,7 +2345,24 @@ async def _create_dataset_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2170,7 +2425,24 @@ async def _create_dataset_version_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2225,7 +2497,24 @@ async def _get_dataset_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Dataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2284,7 +2573,24 @@ async def _get_dataset_version_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetVersion._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2343,7 +2649,24 @@ async def _get_dataset_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DatasetOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2394,7 +2717,24 @@ async def _list_prompts( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListDatasetsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2455,7 +2795,24 @@ async def _list_versions( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListDatasetVersionsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2503,7 +2860,24 @@ async def _delete_dataset( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeletePromptOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2560,7 +2934,24 @@ async def _delete_dataset_version( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeletePromptVersionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2619,7 +3010,24 @@ async def _restore_version( response_dict = {} if not response.body else json.loads(response.body) return_value = types.RestoreVersionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2686,7 +3094,24 @@ async def _update_dataset_resource( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Dataset._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2744,7 +3169,24 @@ async def _create_custom_job_resource( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2799,7 +3241,24 @@ async def _get_custom_job( response_dict = _CustomJob_from_vertex(response_dict) return_value = types.CustomJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -2854,7 +3313,24 @@ async def _optimize( response_dict = {} if not response.body else json.loads(response.body) return_value = types.OptimizeResponseEndpoint._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/sandboxes.py b/vertexai/_genai/sandboxes.py index 6b844848cf..dcea0f260e 100644 --- a/vertexai/_genai/sandboxes.py +++ b/vertexai/_genai/sandboxes.py @@ -215,7 +215,24 @@ def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSandboxOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -276,7 +293,24 @@ def _delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineSandboxOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -333,7 +367,24 @@ def _execute_code( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ExecuteSandboxEnvironmentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -394,7 +445,24 @@ def _get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.SandboxEnvironment._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -459,7 +527,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineSandboxesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -510,7 +595,24 @@ def _get_sandbox_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSandboxOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -915,7 +1017,24 @@ async def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSandboxOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -978,7 +1097,24 @@ async def _delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineSandboxOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1037,7 +1173,24 @@ async def _execute_code( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ExecuteSandboxEnvironmentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1100,7 +1253,24 @@ async def _get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.SandboxEnvironment._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1167,7 +1337,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineSandboxesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1220,7 +1407,24 @@ async def _get_sandbox_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSandboxOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/session_events.py b/vertexai/_genai/session_events.py index ac969fdce2..11d679846f 100644 --- a/vertexai/_genai/session_events.py +++ b/vertexai/_genai/session_events.py @@ -192,7 +192,24 @@ def append( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AppendAgentEngineSessionEventResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -257,7 +274,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineSessionEventsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -365,7 +399,24 @@ async def append( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AppendAgentEngineSessionEventResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -432,7 +483,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListAgentEngineSessionEventsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) diff --git a/vertexai/_genai/sessions.py b/vertexai/_genai/sessions.py index 68f4876dc6..bc284fcd3f 100644 --- a/vertexai/_genai/sessions.py +++ b/vertexai/_genai/sessions.py @@ -273,7 +273,24 @@ def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -338,7 +355,24 @@ def delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -403,7 +437,24 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Session._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -468,7 +519,24 @@ def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListReasoningEnginesSessionsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -519,7 +587,24 @@ def _get_session_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -584,7 +669,24 @@ def _update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -748,7 +850,24 @@ async def _create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -815,7 +934,24 @@ async def delete( response_dict = {} if not response.body else json.loads(response.body) return_value = types.DeleteAgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -882,7 +1018,24 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Session._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -949,7 +1102,24 @@ async def _list( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ListReasoningEnginesSessionsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1002,7 +1172,24 @@ async def _get_session_operation( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value) @@ -1069,7 +1256,24 @@ async def _update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.AgentEngineSessionOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs=( + { + "config": { + "response_schema": getattr( + parameter_model.config, "response_schema", None + ), + "response_json_schema": getattr( + parameter_model.config, "response_json_schema", None + ), + "include_all_fields": getattr( + parameter_model.config, "include_all_fields", None + ), + } + } + if getattr(parameter_model, "config", None) + else {} + ), ) self._api_client._verify_response(return_value)