-
I'll start with a short description, but I can add a ton of details if this isn't a simple answer. I can also successfully run the SparkPi example using the yarn node locally. In the Resource Manager logs, Yarn assigns the submission a single digit application ID ( Any assistance would be greatly appreciated. |
Beta Was this translation helpful? Give feedback.
Replies: 5 comments
-
The log message itself suggests that something went wrong with sending the request to the Ozone Manager. Can you please enable debug level logging for the hadoop root logger in the job, and collect the exception stack trace that will be printed with this same message on debug level by the RetryInvocationHandler? I guess that will tell us more about the exact failure with the request. |
Beta Was this translation helpful? Give feedback.
-
Sorry it took me a while to get back to this discussion. I've increased the logging levels for spark to see why it wasn't getting to the Ozone manager. On the machine where I run the spark command, I am able to run both I'm attaching the extended output of the spark-submit command, and it looks like it may be related to HDDS-6570? Any guidance is greatly appreciated. |
Beta Was this translation helpful? Give feedback.
-
Yup. https://issues.apache.org/jira/browse/HDDS-6926 addresses this issue. |
Beta Was this translation helpful? Give feedback.
-
Thanks so much for the reply. According to the discussion on the Jira ticket, the problem comes from different jars using shaded vs unshaded, and the ticket is marked at closed, which leads me to think that this issue was resolved in newer versions of Ozone. I'm using Spark 3.4.0 (built for use with Hadoop) and Ozone 1.3.0, along with the version of Yarn included with Hadoop 3.3.5, and I'm still getting the error described in above. I'm sorry if I'm dense here, but is there a way to get Spark working with Ozone without recompiling jar files? Thanks again, |
Beta Was this translation helpful? Give feedback.
-
Thanks to @GeorgeJahad comments in the last community call, I was able to get all this working. The key error in my issue was of trying to cast an unshaded variable as shaded. The fix for this was instead of adding $OZONE_HOME/share/ozone/lib/ozone-filesystem-hadoop3-1.3.0.jar into the necessary places for spark and HDFS, I used ozone-filesystem-hadoop3-client-1.3.0.jar. Following this, I did have to re-do my configs to use o3fs instead of ofs, but now my spark jobs are successfully submitted to yarn, and ran across the yarn cluster. I also had to get the log4j jar file added to the spark classpath, but those errors were rather explicit in what was missing. I'd like to thank everyone who helped for their assistance in getting around this issue. |
Beta Was this translation helpful? Give feedback.
Thanks to @GeorgeJahad comments in the last community call, I was able to get all this working.
The key error in my issue was of trying to cast an unshaded variable as shaded. The fix for this was instead of adding $OZONE_HOME/share/ozone/lib/ozone-filesystem-hadoop3-1.3.0.jar into the necessary places for spark and HDFS, I used ozone-filesystem-hadoop3-client-1.3.0.jar.
Following this, I did have to re-do my configs to use o3fs instead of ofs, but now my spark jobs are successfully submitted to yarn, and ran across the yarn cluster. I also had to get the log4j jar file added to the spark classpath, but those errors were rather explicit in what was missing.
I'd like to thank everyone who …