diff --git a/src/sagemaker/processing.py b/src/sagemaker/processing.py index 9a1d8bd431..1866cdff32 100644 --- a/src/sagemaker/processing.py +++ b/src/sagemaker/processing.py @@ -1622,7 +1622,7 @@ def run( # type: ignore[override] :class:`~sagemaker.workflow.pipeline_context.PipelineSession` """ s3_runproc_sh, inputs, job_name = self._pack_and_upload_code( - code, source_dir, dependencies, git_config, job_name, inputs + code, source_dir, dependencies, git_config, job_name, inputs, kms_key ) # Submit a processing job. @@ -1638,7 +1638,9 @@ def run( # type: ignore[override] kms_key=kms_key, ) - def _pack_and_upload_code(self, code, source_dir, dependencies, git_config, job_name, inputs): + def _pack_and_upload_code( + self, code, source_dir, dependencies, git_config, job_name, inputs, kms_key=None + ): """Pack local code bundle and upload to Amazon S3.""" if code.startswith("s3://"): return code, inputs, job_name @@ -1676,6 +1678,7 @@ def _pack_and_upload_code(self, code, source_dir, dependencies, git_config, job_ s3_runproc_sh = S3Uploader.upload_string_as_file_body( self._generate_framework_script(script), desired_s3_uri=entrypoint_s3_uri, + kms_key=kms_key, sagemaker_session=self.sagemaker_session, ) logger.info("runproc.sh uploaded to %s", s3_runproc_sh) diff --git a/tests/unit/sagemaker/workflow/test_pipeline_session.py b/tests/unit/sagemaker/workflow/test_pipeline_session.py index eca3892390..a9a6fb41c5 100644 --- a/tests/unit/sagemaker/workflow/test_pipeline_session.py +++ b/tests/unit/sagemaker/workflow/test_pipeline_session.py @@ -50,7 +50,7 @@ def test_pipeline_session_init(sagemaker_client_config, boto_session): sagemaker_client=sagemaker_client, ) assert sess.sagemaker_client is not None - assert sess.default_bucket() is not None + assert sess.default_bucket is not None assert sess.context is None