Skip to content

passing configuration for spark processing job #3732

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
ajaiswalgit opened this issue Mar 19, 2023 · 2 comments
Closed

passing configuration for spark processing job #3732

ajaiswalgit opened this issue Mar 19, 2023 · 2 comments

Comments

@ajaiswalgit
Copy link

Describe the feature you'd like
Feature to create spark.processing configuration.json file from kwargs to an S3 path. Many organizations do not have permission to write a file at bucket level.

How would this feature be used? Please describe.
We need to pass few Spark configurations in configurations.json to override default behavior. Configs pass through kwargs.configuration creates a file configuration.json at default bucket level. But user do not have write access at bucket level.

Describe alternatives you've considered
If we can pass default s3 path along with default S3 bucket then it will not fail.

Additional context
User should be able to pass user_defined_s3_folder so that configuration.json is created at S3bucket/user_defined_s3_folder where they have got write permission.

s3_uri = (
f"s3://{self.sagemaker_session.default_bucket()}/{user_defined_s3_folder}/{self._current_job_name}/"
f"input/{self._conf_container_input_name}/{self._conf_file_name}"
)

@jmahlik
Copy link
Contributor

jmahlik commented Aug 23, 2023

Looks like this might be related to #3200 and possibly fixed.

@martinRenou
Copy link
Collaborator

Closing as fixed by #3486

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants