@@ -128,7 +128,6 @@ def _build_for_torchserve(self) -> Model:
128128 Returns:
129129 Model: Configured model ready for TorchServe deployment.
130130 """
131- self .secret_key = ""
132131
133132 # Save inference spec if we have local artifacts
134133 self ._save_model_inference_spec ()
@@ -150,7 +149,7 @@ def _build_for_torchserve(self) -> Model:
150149
151150 # Prepare TorchServe artifacts for local container mode
152151 if self .mode == Mode .LOCAL_CONTAINER and self .model_path :
153- self . secret_key = prepare_for_torchserve (
152+ prepare_for_torchserve (
154153 model_path = self .model_path ,
155154 shared_libs = self .shared_libs ,
156155 dependencies = self .dependencies ,
@@ -159,7 +158,7 @@ def _build_for_torchserve(self) -> Model:
159158 inference_spec = self .inference_spec ,
160159 )
161160 if self .mode == Mode .SAGEMAKER_ENDPOINT and self .model_path :
162- self . secret_key = prepare_for_torchserve (
161+ prepare_for_torchserve (
163162 model_path = self .model_path ,
164163 shared_libs = self .shared_libs ,
165164 dependencies = self .dependencies ,
@@ -187,7 +186,6 @@ def _build_for_tgi(self) -> Model:
187186 Returns:
188187 Model: Configured model ready for TGI deployment.
189188 """
190- self .secret_key = ""
191189
192190 # Initialize TGI-specific configuration
193191 if self .model_server != ModelServer .TGI :
@@ -299,7 +297,6 @@ def _build_for_djl(self) -> Model:
299297 Returns:
300298 Model: Configured model ready for DJL Serving deployment.
301299 """
302- self .secret_key = ""
303300 self .model_server = ModelServer .DJL_SERVING
304301
305302 # Set MODEL_LOADING_TIMEOUT from instance variable
@@ -408,7 +405,6 @@ def _build_for_triton(self) -> Model:
408405 Returns:
409406 Model: Configured model ready for Triton deployment.
410407 """
411- self .secret_key = ""
412408 self ._validate_for_triton ()
413409
414410 if isinstance (self .model , str ):
@@ -467,7 +463,6 @@ def _build_for_tensorflow_serving(self) -> Model:
467463 Raises:
468464 ValueError: If image_uri is not provided for TensorFlow Serving.
469465 """
470- self .secret_key = ""
471466 if not getattr (self , "_is_mlflow_model" , False ):
472467 raise ValueError ("Tensorflow Serving is currently only supported for mlflow models." )
473468
@@ -481,7 +476,7 @@ def _build_for_tensorflow_serving(self) -> Model:
481476 raise ValueError ("image_uri is required for TensorFlow Serving deployment" )
482477
483478 # Prepare TensorFlow Serving artifacts for local container mode
484- self . secret_key = prepare_for_tf_serving (
479+ prepare_for_tf_serving (
485480 model_path = self .model_path ,
486481 shared_libs = self .shared_libs ,
487482 dependencies = self .dependencies ,
@@ -506,7 +501,6 @@ def _build_for_tei(self) -> Model:
506501 Returns:
507502 Model: Configured model ready for TEI deployment.
508503 """
509- self .secret_key = ""
510504
511505 # Set MODEL_LOADING_TIMEOUT from instance variable
512506 if self .model_data_download_timeout :
@@ -592,7 +586,6 @@ def _build_for_smd(self) -> Model:
592586 Returns:
593587 Model: Configured model ready for SMD deployment.
594588 """
595- self .secret_key = ""
596589
597590 self ._save_model_inference_spec ()
598591
@@ -602,7 +595,7 @@ def _build_for_smd(self) -> Model:
602595 cpu_or_gpu = self ._get_processing_unit ()
603596 self .image_uri = self ._get_smd_image_uri (processing_unit = cpu_or_gpu )
604597
605- self . secret_key = prepare_for_smd (
598+ prepare_for_smd (
606599 model_path = self .model_path ,
607600 shared_libs = self .shared_libs ,
608601 dependencies = self .dependencies ,
@@ -626,7 +619,6 @@ def _build_for_transformers(self) -> Model:
626619 Returns:
627620 Model: Configured model ready for Transformers deployment.
628621 """
629- self .secret_key = ""
630622 self .model_server = ModelServer .MMS
631623
632624 # Set MODEL_LOADING_TIMEOUT from instance variable
@@ -646,7 +638,7 @@ def _build_for_transformers(self) -> Model:
646638 self ._create_conda_env ()
647639
648640 if self .mode in [Mode .LOCAL_CONTAINER ] and self .model_path :
649- self . secret_key = prepare_for_mms (
641+ prepare_for_mms (
650642 model_path = self .model_path ,
651643 shared_libs = self .shared_libs ,
652644 dependencies = self .dependencies ,
@@ -655,7 +647,7 @@ def _build_for_transformers(self) -> Model:
655647 inference_spec = self .inference_spec ,
656648 )
657649 if self .mode == Mode .SAGEMAKER_ENDPOINT and self .model_path :
658- self . secret_key = prepare_for_mms (
650+ prepare_for_mms (
659651 model_path = self .model_path ,
660652 shared_libs = self .shared_libs ,
661653 dependencies = self .dependencies ,
@@ -725,7 +717,6 @@ def _build_for_djl_jumpstart(self, init_kwargs) -> Model:
725717 Returns:
726718 Model: Configured DJL model for JumpStart deployment.
727719 """
728- self .secret_key = ""
729720 self .model_server = ModelServer .DJL_SERVING
730721
731722 from sagemaker .serve .model_server .djl_serving .prepare import _create_dir_structure
@@ -761,7 +752,6 @@ def _build_for_tgi_jumpstart(self, init_kwargs) -> Model:
761752 Returns:
762753 Model: Configured TGI model for JumpStart deployment.
763754 """
764- self .secret_key = ""
765755 self .model_server = ModelServer .TGI
766756
767757 from sagemaker .serve .model_server .tgi .prepare import _create_dir_structure
@@ -797,7 +787,6 @@ def _build_for_mms_jumpstart(self, init_kwargs) -> Model:
797787 Returns:
798788 Model: Configured MMS model for JumpStart deployment.
799789 """
800- self .secret_key = ""
801790 self .model_server = ModelServer .MMS
802791
803792 from sagemaker .serve .model_server .multi_model_server .prepare import _create_dir_structure
@@ -840,7 +829,6 @@ def _build_for_jumpstart(self) -> Model:
840829 """
841830 from sagemaker .core .jumpstart .factory .utils import get_init_kwargs
842831
843- self .secret_key = ""
844832
845833 # Get JumpStart model configuration
846834 init_kwargs = get_init_kwargs (
0 commit comments