Skip to content

Commit 9c0af74

Browse files
wangyumsrowen
authored andcommitted
[SPARK-27175][BUILD] Upgrade hadoop-3 to 3.2.0
## What changes were proposed in this pull request? This PR upgrade `hadoop-3` to `3.2.0` to workaround [HADOOP-16086](https://issues.apache.org/jira/browse/HADOOP-16086). Otherwise some test case will throw IllegalArgumentException: ```java 02:44:34.707 ERROR org.apache.hadoop.hive.ql.exec.Task: Job Submission failed with exception 'java.io.IOException(Cannot initialize Cluster. Please check your configuration for mapreduce.framework.name and the correspond server addresses.)' java.io.IOException: Cannot initialize Cluster. Please check your configuration for mapreduce.framework.name and the correspond server addresses. at org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:116) at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:109) at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:102) at org.apache.hadoop.mapred.JobClient.init(JobClient.java:475) at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:454) at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:369) at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:151) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227) at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$runHive$1(HiveClientImpl.scala:730) at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:283) at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:221) at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:220) at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:266) at org.apache.spark.sql.hive.client.HiveClientImpl.runHive(HiveClientImpl.scala:719) at org.apache.spark.sql.hive.client.HiveClientImpl.runSqlHive(HiveClientImpl.scala:709) at org.apache.spark.sql.hive.StatisticsSuite.createNonPartitionedTable(StatisticsSuite.scala:719) at org.apache.spark.sql.hive.StatisticsSuite.$anonfun$testAlterTableProperties$2(StatisticsSuite.scala:822) ``` ## How was this patch tested? manual tests Closes apache#24106 from wangyum/SPARK-27175. Authored-by: Yuming Wang <[email protected]> Signed-off-by: Sean Owen <[email protected]>
1 parent 4adbcdc commit 9c0af74

File tree

6 files changed

+31
-29
lines changed

6 files changed

+31
-29
lines changed

dev/deps/spark-deps-hadoop-3.1 dev/deps/spark-deps-hadoop-3.2

+22-20
Original file line numberDiff line numberDiff line change
@@ -15,6 +15,7 @@ arpack_combined_all-0.1.jar
1515
arrow-format-0.12.0.jar
1616
arrow-memory-0.12.0.jar
1717
arrow-vector-0.12.0.jar
18+
audience-annotations-0.5.0.jar
1819
automaton-1.11-8.jar
1920
avro-1.8.2.jar
2021
avro-ipc-1.8.2.jar
@@ -42,11 +43,12 @@ commons-logging-1.1.3.jar
4243
commons-math3-3.4.1.jar
4344
commons-net-3.1.jar
4445
commons-pool-1.5.4.jar
46+
commons-text-1.6.jar
4547
compress-lzf-1.0.3.jar
4648
core-1.1.2.jar
47-
curator-client-2.12.0.jar
48-
curator-framework-2.12.0.jar
49-
curator-recipes-2.12.0.jar
49+
curator-client-2.13.0.jar
50+
curator-framework-2.13.0.jar
51+
curator-recipes-2.13.0.jar
5052
datanucleus-api-jdo-3.2.6.jar
5153
datanucleus-core-3.2.10.jar
5254
datanucleus-rdbms-3.2.9.jar
@@ -60,20 +62,20 @@ gson-2.2.4.jar
6062
guava-14.0.1.jar
6163
guice-4.0.jar
6264
guice-servlet-4.0.jar
63-
hadoop-annotations-3.1.0.jar
64-
hadoop-auth-3.1.0.jar
65-
hadoop-client-3.1.0.jar
66-
hadoop-common-3.1.0.jar
67-
hadoop-hdfs-client-3.1.0.jar
68-
hadoop-mapreduce-client-common-3.1.0.jar
69-
hadoop-mapreduce-client-core-3.1.0.jar
70-
hadoop-mapreduce-client-jobclient-3.1.0.jar
71-
hadoop-yarn-api-3.1.0.jar
72-
hadoop-yarn-client-3.1.0.jar
73-
hadoop-yarn-common-3.1.0.jar
74-
hadoop-yarn-registry-3.1.0.jar
75-
hadoop-yarn-server-common-3.1.0.jar
76-
hadoop-yarn-server-web-proxy-3.1.0.jar
65+
hadoop-annotations-3.2.0.jar
66+
hadoop-auth-3.2.0.jar
67+
hadoop-client-3.2.0.jar
68+
hadoop-common-3.2.0.jar
69+
hadoop-hdfs-client-3.2.0.jar
70+
hadoop-mapreduce-client-common-3.2.0.jar
71+
hadoop-mapreduce-client-core-3.2.0.jar
72+
hadoop-mapreduce-client-jobclient-3.2.0.jar
73+
hadoop-yarn-api-3.2.0.jar
74+
hadoop-yarn-client-3.2.0.jar
75+
hadoop-yarn-common-3.2.0.jar
76+
hadoop-yarn-registry-3.2.0.jar
77+
hadoop-yarn-server-common-3.2.0.jar
78+
hadoop-yarn-server-web-proxy-3.2.0.jar
7779
hk2-api-2.4.0-b34.jar
7880
hk2-locator-2.4.0-b34.jar
7981
hk2-utils-2.4.0-b34.jar
@@ -88,8 +90,8 @@ jackson-core-2.9.8.jar
8890
jackson-core-asl-1.9.13.jar
8991
jackson-databind-2.9.8.jar
9092
jackson-dataformat-yaml-2.9.8.jar
91-
jackson-jaxrs-base-2.7.8.jar
92-
jackson-jaxrs-json-provider-2.7.8.jar
93+
jackson-jaxrs-base-2.9.5.jar
94+
jackson-jaxrs-json-provider-2.9.5.jar
9395
jackson-mapper-asl-1.9.13.jar
9496
jackson-module-jaxb-annotations-2.9.8.jar
9597
jackson-module-paranamer-2.9.8.jar
@@ -215,5 +217,5 @@ woodstox-core-5.0.3.jar
215217
xbean-asm7-shaded-4.12.jar
216218
xz-1.5.jar
217219
zjsonpatch-0.3.0.jar
218-
zookeeper-3.4.9.jar
220+
zookeeper-3.4.13.jar
219221
zstd-jni-1.3.2-2.jar

dev/run-tests-jenkins.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -173,8 +173,8 @@ def main():
173173
os.environ["AMPLAB_JENKINS_BUILD_PROFILE"] = "hadoop2.6"
174174
if "test-hadoop2.7" in ghprb_pull_title:
175175
os.environ["AMPLAB_JENKINS_BUILD_PROFILE"] = "hadoop2.7"
176-
if "test-hadoop3.1" in ghprb_pull_title:
177-
os.environ["AMPLAB_JENKINS_BUILD_PROFILE"] = "hadoop3.1"
176+
if "test-hadoop3.2" in ghprb_pull_title:
177+
os.environ["AMPLAB_JENKINS_BUILD_PROFILE"] = "hadoop3.2"
178178

179179
build_display_name = os.environ["BUILD_DISPLAY_NAME"]
180180
build_url = os.environ["BUILD_URL"]

dev/run-tests.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -280,7 +280,7 @@ def get_hadoop_profiles(hadoop_version):
280280

281281
sbt_maven_hadoop_profiles = {
282282
"hadoop2.7": ["-Phadoop-2.7"],
283-
"hadoop3.1": ["-Phadoop-3.1"],
283+
"hadoop3.2": ["-Phadoop-3.2"],
284284
}
285285

286286
if hadoop_version in sbt_maven_hadoop_profiles:

dev/test-dependencies.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ HADOOP2_MODULE_PROFILES="-Phive-thriftserver -Pmesos -Pkubernetes -Pyarn -Phive"
3333
MVN="build/mvn"
3434
HADOOP_PROFILES=(
3535
hadoop-2.7
36-
hadoop-3.1
36+
hadoop-3.2
3737
)
3838

3939
# We'll switch the version to a temp. one, publish POMs using that new version, then switch back to

hadoop-cloud/pom.xml

+1-1
Original file line numberDiff line numberDiff line change
@@ -199,7 +199,7 @@
199199
enables store-specific committers.
200200
-->
201201
<profile>
202-
<id>hadoop-3.1</id>
202+
<id>hadoop-3.2</id>
203203
<dependencies>
204204
<!--
205205
There's now a hadoop-cloud-storage which transitively pulls in the store JARs,

pom.xml

+4-4
Original file line numberDiff line numberDiff line change
@@ -2646,11 +2646,11 @@
26462646
</profile>
26472647

26482648
<profile>
2649-
<id>hadoop-3.1</id>
2649+
<id>hadoop-3.2</id>
26502650
<properties>
2651-
<hadoop.version>3.1.0</hadoop.version>
2652-
<curator.version>2.12.0</curator.version>
2653-
<zookeeper.version>3.4.9</zookeeper.version>
2651+
<hadoop.version>3.2.0</hadoop.version>
2652+
<curator.version>2.13.0</curator.version>
2653+
<zookeeper.version>3.4.13</zookeeper.version>
26542654
</properties>
26552655
</profile>
26562656

0 commit comments

Comments
 (0)