-
Notifications
You must be signed in to change notification settings - Fork 28.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SPARK-48651][DOC] Configuring different JDK for Spark on YARN #47010
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
--conf spark.yarn.appMasterEnv.JAVA_HOME=./openjdk-21.tar.gz/openjdk-21 \ | ||
--conf spark.executorEnv.JAVA_HOME=./openjdk-21.tar.gz/openjdk-21 \ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@yaooqinn @tgravescs sorry for correcting this in 5bbe200 after your approval, I also updated the PR description to add the manual test result on a YARN cluster
thanks, merging to master! |
What changes were proposed in this pull request?
This PR updates the Spark on YARN docs to guide users to configure a different JDK for Spark Applications.
Why are the changes needed?
As of today, the latest Apache Hadoop 3.4.0 does not support Java 17 yet, while Spark 4.0.0 requires at least Java 17, so users who want to use Spark on YARN must configure a different JDK for Spark applications run on YARN.
This is also asked in the mailing list https://lists.apache.org/thread/ply807h0hht1h8o7x7g1s3j51mnot5dr
Does this PR introduce any user-facing change?
Yes, it changes the user docs.
How was this patch tested?
I verified the command in a YARN cluster.
The following command submits a Spark application with the distributed JDK 21
Was this patch authored or co-authored using generative AI tooling?
No.