From eca15e1a70da6b7bfd91eeac0d745285800ab843 Mon Sep 17 00:00:00 2001 From: Ayush Agrawal Date: Sat, 4 Apr 2026 19:19:28 -0700 Subject: [PATCH] chore: update SDK response serialization logic PiperOrigin-RevId: 894725483 --- google/genai/batches.py | 170 +++++++++- google/genai/caches.py | 170 +++++++++- google/genai/documents.py | 68 +++- google/genai/file_search_stores.py | 170 +++++++++- google/genai/files.py | 170 +++++++++- google/genai/models.py | 510 +++++++++++++++++++++++++++-- google/genai/operations.py | 34 +- google/genai/tunings.py | 170 +++++++++- 8 files changed, 1376 insertions(+), 86 deletions(-) diff --git a/google/genai/batches.py b/google/genai/batches.py index 0dbd0b8ac..ad26c963d 100644 --- a/google/genai/batches.py +++ b/google/genai/batches.py @@ -1650,7 +1650,22 @@ def _create( response_dict = _BatchJob_from_mldev(response_dict) return_value = types.BatchJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1710,7 +1725,22 @@ def _create_embeddings( response_dict = _BatchJob_from_mldev(response_dict) return_value = types.BatchJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1790,7 +1820,22 @@ def get( response_dict = _BatchJob_from_mldev(response_dict) return_value = types.BatchJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1914,7 +1959,22 @@ def _list( response_dict = _ListBatchJobsResponse_from_mldev(response_dict) return_value = types.ListBatchJobsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1999,7 +2059,22 @@ def delete( response_dict = _DeleteResourceJob_from_mldev(response_dict) return_value = types.DeleteResourceJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -2195,7 +2270,22 @@ async def _create( response_dict = _BatchJob_from_mldev(response_dict) return_value = types.BatchJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -2255,7 +2345,22 @@ async def _create_embeddings( response_dict = _BatchJob_from_mldev(response_dict) return_value = types.BatchJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -2337,7 +2442,22 @@ async def get( response_dict = _BatchJob_from_mldev(response_dict) return_value = types.BatchJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -2463,7 +2583,22 @@ async def _list( response_dict = _ListBatchJobsResponse_from_mldev(response_dict) return_value = types.ListBatchJobsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -2548,7 +2683,22 @@ async def delete( response_dict = _DeleteResourceJob_from_mldev(response_dict) return_value = types.DeleteResourceJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers diff --git a/google/genai/caches.py b/google/genai/caches.py index 9ba17239d..e4263d09d 100644 --- a/google/genai/caches.py +++ b/google/genai/caches.py @@ -1105,7 +1105,22 @@ def create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.CachedContent._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1170,7 +1185,22 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.CachedContent._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1245,7 +1275,22 @@ def delete( response_dict = _DeleteCachedContentResponse_from_mldev(response_dict) return_value = types.DeleteCachedContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1319,7 +1364,22 @@ def update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.CachedContent._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1375,7 +1435,22 @@ def _list( response_dict = _ListCachedContentsResponse_from_mldev(response_dict) return_value = types.ListCachedContentsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1487,7 +1562,22 @@ async def create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.CachedContent._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1555,7 +1645,22 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.CachedContent._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1631,7 +1736,22 @@ async def delete( response_dict = _DeleteCachedContentResponse_from_mldev(response_dict) return_value = types.DeleteCachedContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1705,7 +1825,22 @@ async def update( response_dict = {} if not response.body else json.loads(response.body) return_value = types.CachedContent._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1763,7 +1898,22 @@ async def _list( response_dict = _ListCachedContentsResponse_from_mldev(response_dict) return_value = types.ListCachedContentsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers diff --git a/google/genai/documents.py b/google/genai/documents.py index 0e7016156..5c7a77f0e 100644 --- a/google/genai/documents.py +++ b/google/genai/documents.py @@ -182,7 +182,22 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Document._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -291,7 +306,22 @@ def _list( response_dict = _ListDocumentsResponse_from_mldev(response_dict) return_value = types.ListDocumentsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -384,7 +414,22 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.Document._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -495,7 +540,22 @@ async def _list( response_dict = _ListDocumentsResponse_from_mldev(response_dict) return_value = types.ListDocumentsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) diff --git a/google/genai/file_search_stores.py b/google/genai/file_search_stores.py index 4a898c6b4..fd40b000b 100644 --- a/google/genai/file_search_stores.py +++ b/google/genai/file_search_stores.py @@ -371,7 +371,22 @@ def create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.FileSearchStore._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -434,7 +449,22 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.FileSearchStore._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -540,7 +570,22 @@ def _list( response_dict = _ListFileSearchStoresResponse_from_mldev(response_dict) return_value = types.ListFileSearchStoresResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -613,7 +658,22 @@ def _upload_to_file_search_store( return_value = ( types.UploadToFileSearchStoreResumableResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) ) @@ -689,7 +749,22 @@ def import_file( response_dict = _ImportFileOperation_from_mldev(response_dict) return_value = types.ImportFileOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -861,7 +936,22 @@ async def create( response_dict = {} if not response.body else json.loads(response.body) return_value = types.FileSearchStore._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -926,7 +1016,22 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.FileSearchStore._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1034,7 +1139,22 @@ async def _list( response_dict = _ListFileSearchStoresResponse_from_mldev(response_dict) return_value = types.ListFileSearchStoresResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -1107,7 +1227,22 @@ async def _upload_to_file_search_store( return_value = ( types.UploadToFileSearchStoreResumableResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) ) @@ -1183,7 +1318,22 @@ async def import_file( response_dict = _ImportFileOperation_from_mldev(response_dict) return_value = types.ImportFileOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) diff --git a/google/genai/files.py b/google/genai/files.py index 6bbbd65b0..7e09b87d3 100644 --- a/google/genai/files.py +++ b/google/genai/files.py @@ -225,7 +225,22 @@ def _list( response_dict = _ListFilesResponse_from_mldev(response_dict) return_value = types.ListFilesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -290,7 +305,22 @@ def _create( response_dict = _CreateFileResponse_from_mldev(response_dict) return_value = types.CreateFileResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -355,7 +385,22 @@ def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.File._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -424,7 +469,22 @@ def delete( response_dict = _DeleteFileResponse_from_mldev(response_dict) return_value = types.DeleteFileResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -489,7 +549,22 @@ def _register_files( response_dict = _RegisterFilesResponse_from_mldev(response_dict) return_value = types.RegisterFilesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -762,7 +837,22 @@ async def _list( response_dict = _ListFilesResponse_from_mldev(response_dict) return_value = types.ListFilesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -827,7 +917,22 @@ async def _create( response_dict = _CreateFileResponse_from_mldev(response_dict) return_value = types.CreateFileResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -894,7 +999,22 @@ async def get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.File._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -963,7 +1083,22 @@ async def delete( response_dict = _DeleteFileResponse_from_mldev(response_dict) return_value = types.DeleteFileResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1028,7 +1163,22 @@ async def _register_files( response_dict = _RegisterFilesResponse_from_mldev(response_dict) return_value = types.RegisterFilesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) diff --git a/google/genai/models.py b/google/genai/models.py index d948b0dbd..aa63ff4a9 100644 --- a/google/genai/models.py +++ b/google/genai/models.py @@ -4704,7 +4704,22 @@ def _generate_content( ) return_value = types.GenerateContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -4790,7 +4805,22 @@ def _generate_content_stream( ) return_value = types.GenerateContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -4896,7 +4926,22 @@ def _embed_content( ) return_value = types.EmbedContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -4972,7 +5017,22 @@ def _generate_images( ) return_value = types.GenerateImagesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -5038,7 +5098,22 @@ def _edit_image( ) return_value = types.EditImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -5104,7 +5179,22 @@ def _upscale_image( ) return_value = types.UpscaleImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -5214,7 +5304,22 @@ def recontext_image( ) return_value = types.RecontextImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -5299,7 +5404,22 @@ def segment_image( ) return_value = types.SegmentImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -5360,7 +5480,22 @@ def get( response_dict = _Model_from_mldev(response_dict, None, parameter_model) return_value = types.Model._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -5429,7 +5564,22 @@ def _list( ) return_value = types.ListModelsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -5497,7 +5647,22 @@ def update( response_dict = _Model_from_mldev(response_dict, None, parameter_model) return_value = types.Model._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -5567,7 +5732,22 @@ def delete( ) return_value = types.DeleteModelResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -5662,7 +5842,22 @@ def count_tokens( ) return_value = types.CountTokensResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -5747,7 +5942,22 @@ def compute_tokens( ) return_value = types.ComputeTokensResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -5829,7 +6039,22 @@ def _generate_videos( ) return_value = types.GenerateVideosOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -6663,7 +6888,22 @@ async def _generate_content( ) return_value = types.GenerateContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -6752,7 +6992,22 @@ async def async_generator(): # type: ignore[no-untyped-def] ) return_value = types.GenerateContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -6860,7 +7115,22 @@ async def _embed_content( ) return_value = types.EmbedContentResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -6936,7 +7206,22 @@ async def _generate_images( ) return_value = types.GenerateImagesResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -7002,7 +7287,22 @@ async def _edit_image( ) return_value = types.EditImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -7068,7 +7368,22 @@ async def _upscale_image( ) return_value = types.UpscaleImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -7178,7 +7493,22 @@ async def recontext_image( ) return_value = types.RecontextImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -7266,7 +7596,22 @@ async def segment_image( ) return_value = types.SegmentImageResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -7329,7 +7674,22 @@ async def get( response_dict = _Model_from_mldev(response_dict, None, parameter_model) return_value = types.Model._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -7400,7 +7760,22 @@ async def _list( ) return_value = types.ListModelsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -7468,7 +7843,22 @@ async def update( response_dict = _Model_from_mldev(response_dict, None, parameter_model) return_value = types.Model._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -7538,7 +7928,22 @@ async def delete( ) return_value = types.DeleteModelResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -7633,7 +8038,22 @@ async def count_tokens( ) return_value = types.CountTokensResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -7717,7 +8137,22 @@ async def compute_tokens( ) return_value = types.ComputeTokensResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -7799,7 +8234,22 @@ async def _generate_videos( ) return_value = types.GenerateVideosOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) diff --git a/google/genai/operations.py b/google/genai/operations.py index fdff3933c..8a0de8c32 100644 --- a/google/genai/operations.py +++ b/google/genai/operations.py @@ -235,7 +235,22 @@ def _get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ProjectOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) @@ -473,7 +488,22 @@ async def _get( response_dict = {} if not response.body else json.loads(response.body) return_value = types.ProjectOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) self._api_client._verify_response(return_value) diff --git a/google/genai/tunings.py b/google/genai/tunings.py index 6afb36a8a..bdf93c186 100644 --- a/google/genai/tunings.py +++ b/google/genai/tunings.py @@ -1602,7 +1602,22 @@ def _get( ) return_value = types.TuningJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1668,7 +1683,22 @@ def _list( ) return_value = types.ListTuningJobsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1746,7 +1776,22 @@ def cancel( ) return_value = types.CancelTuningJobResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1821,7 +1866,22 @@ def _tune( ) return_value = types.TuningJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -1898,7 +1958,22 @@ def _tune_mldev( ) return_value = types.TuningOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -2110,7 +2185,22 @@ async def _get( ) return_value = types.TuningJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -2178,7 +2268,22 @@ async def _list( ) return_value = types.ListTuningJobsResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -2256,7 +2361,22 @@ async def cancel( ) return_value = types.CancelTuningJobResponse._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -2331,7 +2451,22 @@ async def _tune( ) return_value = types.TuningJob._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers @@ -2408,7 +2543,22 @@ async def _tune_mldev( ) return_value = types.TuningOperation._from_response( - response=response_dict, kwargs=parameter_model.model_dump() + response=response_dict, + kwargs={ + 'config': { + 'response_schema': getattr( + parameter_model.config, 'response_schema', None + ), + 'response_json_schema': getattr( + parameter_model.config, 'response_json_schema', None + ), + 'include_all_fields': getattr( + parameter_model.config, 'include_all_fields', None + ), + } + } + if getattr(parameter_model, 'config', None) + else {}, ) return_value.sdk_http_response = types.HttpResponse( headers=response.headers