Skip to content

Commit e347827

Browse files
authored
Remove Spark 2.x (#2316)
1 parent f277311 commit e347827

File tree

101 files changed

+28
-11391
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

101 files changed

+28
-11391
lines changed

dist/build.gradle

+9-9
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ base {
1212
archivesName = 'elasticsearch-hadoop'
1313
}
1414

15-
def sparkVariantIncluded = 'spark20scala211'
15+
def sparkVariantIncluded = 'spark30scala212'
1616

1717
configurations {
1818
embedded {
@@ -47,17 +47,17 @@ configurations {
4747

4848
BuildPlugin.disableTransitiveDependencies(project, project.configurations.thirdPartyShaded)
4949

50-
def distProjects = [":elasticsearch-hadoop-mr", ":elasticsearch-hadoop-hive",
51-
":elasticsearch-spark-20"]
50+
def distProjects = [":elasticsearch-hadoop-mr", ":elasticsearch-hadoop-hive", ":elasticsearch-spark-30"]
5251

5352
distProjects.each { distProject ->
5453
def configureDistDependency = { Dependency dependency ->
55-
if (distProject == ":elasticsearch-spark-20") {
54+
if (distProject == ":elasticsearch-spark-30") {
5655
dependency.capabilities {
5756
requireCapability("org.elasticsearch.spark.sql.variant:$sparkVariantIncluded:$project.version")
5857
}
5958
}
6059
}
60+
6161
dependencies {
6262
// This is only going to pull in each project's regular jar to create the project-wide uberjar.
6363
add('embedded', project(distProject), configureDistDependency)
@@ -86,25 +86,25 @@ dependencies {
8686
}
8787
implementation("org.apache.hive:hive-exec:$hiveVersion")
8888
implementation("org.apache.hive:hive-metastore:$hiveVersion")
89-
implementation("org.apache.spark:spark-core_${project.ext.scala211MajorVersion}:$spark20Version") {
89+
implementation("org.apache.spark:spark-core_${project.ext.scala212MajorVersion}:$spark30Version") {
9090
exclude group: 'javax.servlet'
9191
exclude group: 'org.apache.hadoop'
9292
}
93-
implementation("org.apache.spark:spark-yarn_${project.ext.scala211MajorVersion}:$spark20Version") {
93+
implementation("org.apache.spark:spark-yarn_${project.ext.scala212MajorVersion}:$spark30Version") {
9494
exclude group: 'org.apache.hadoop'
9595
}
96-
implementation("org.apache.spark:spark-sql_${project.ext.scala211MajorVersion}:$spark20Version") {
96+
implementation("org.apache.spark:spark-sql_${project.ext.scala212MajorVersion}:$spark30Version") {
9797
exclude group: 'org.apache.hadoop'
9898
}
99-
implementation("org.apache.spark:spark-streaming_${project.ext.scala211MajorVersion}:$spark20Version") {
99+
implementation("org.apache.spark:spark-streaming_${project.ext.scala212MajorVersion}:$spark30Version") {
100100
exclude group: 'org.apache.hadoop'
101101
}
102102
implementation("org.scala-lang:scala-library:$scala211Version")
103103
implementation("org.scala-lang:scala-reflect:$scala211Version")
104104
implementation(project.ext.hadoopClient)
105105
implementation("org.apache.hadoop:hadoop-common:${project.ext.hadoopVersion}")
106106
implementation("org.apache.hadoop:hadoop-mapreduce-client-core:${project.ext.hadoopVersion}")
107-
compileOnly("org.apache.spark:spark-catalyst_${project.ext.scala211MajorVersion}:$spark20Version")
107+
compileOnly("org.apache.spark:spark-catalyst_${project.ext.scala212MajorVersion}:$spark30Version")
108108
}
109109

110110
// Configure uber jar

dist/licenses/spark-core_2.11-2.3.0.jar.sha1

-1
This file was deleted.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
6f7105e792cbb650eca44890f0444720bd6a8204

dist/licenses/spark-sql_2.11-2.3.0.jar.sha1

-1
This file was deleted.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
a28ead1aa19899654296a6f8a458f8f207f89a73

dist/licenses/spark-streaming_2.11-2.3.0.jar.sha1

-1
This file was deleted.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
e4a9f76eff1db09c67f855cdf60cd83bbfad69ab

dist/licenses/spark-yarn_2.11-2.3.0.jar.sha1

-1
This file was deleted.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
0c8dd1e45fbc589d2438a889c7fe98f0e9fd77ec

docs/src/reference/asciidoc/appendix/breaking.adoc

+10
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,16 @@ For clarity, we always list any breaking changes at the top of the
88
//NOTE: The notable-breaking-changes tagged regions are re-used in the
99
//Installation and Upgrade Guide
1010

11+
[[breaking-changes-90]]
12+
=== Breaking Changes in 9.0
13+
14+
This section details the breaking changes when upgrading {eh} from 8.x to 9.0.
15+
16+
[[removals-9.0]]
17+
==== Removal of Spark 2.x
18+
19+
Support for the Spark 2.x has been removed in {eh} 9.0. Spark 3.x is the new default supported version.
20+
1121
=== Deprecations in 8.18
1222

1323
The following functionality has been deprecated in {eh} 8.18 and will be removed

docs/src/reference/asciidoc/core/intro/download.adoc

+2-2
Original file line numberDiff line numberDiff line change
@@ -69,8 +69,8 @@ The Spark connector framework is the most sensitive to version incompatibilities
6969
| Spark Version | Scala Version | ES-Hadoop Artifact ID
7070
| 1.0 - 2.x | 2.10 | <unsupported>
7171
| 1.0 - 1.6 | 2.11 | <unsupported>
72-
| 2.x | 2.11 | elasticsearch-spark-20_2.11
73-
| 2.x | 2.12 | elasticsearch-spark-20_2.12
72+
| 2.x | 2.11 | <unsupported>
73+
| 2.x | 2.12 | <unsupported>
7474
| 3.0+ | 2.12 | elasticsearch-spark-30_2.12
7575
| 3.2+ | 2.13 | elasticsearch-spark-30_2.13
7676
|==========================================================

settings.gradle

-4
Original file line numberDiff line numberDiff line change
@@ -21,10 +21,6 @@ include 'spark-core'
2121
project(":spark-core").projectDir = new File(settingsDir, "spark/core")
2222
project(":spark-core").name = "elasticsearch-spark"
2323

24-
include 'sql-20'
25-
project(":sql-20").projectDir = new File(settingsDir, "spark/sql-20")
26-
project(":sql-20").name = "elasticsearch-spark-20"
27-
2824
include 'sql-30'
2925
project(":sql-30").projectDir = new File(settingsDir, "spark/sql-30")
3026
project(":sql-30").name = "elasticsearch-spark-30"

spark/core/build.gradle

+2-38
Original file line numberDiff line numberDiff line change
@@ -10,12 +10,10 @@ apply plugin: 'spark.variants'
1010
sparkVariants {
1111
capabilityGroup 'org.elasticsearch.spark.variant'
1212

13-
// Changing the formatting of these lines could break .buildkite/pipeline.py, it uses regex to parse the `spark20scala212` part
13+
// Changing the formatting of these lines could break .buildkite/pipeline.py, it uses regex to parse the `spark30scala212` part
1414
// We should maybe move these to a separate config file that can be read from both this file and the pipeline script in the future if it creates issues
15-
setCoreDefaultVariant "spark20scala212", spark24Version, scala212Version
15+
setCoreDefaultVariant "spark30scala212", spark30Version, scala212Version
1616
addCoreFeatureVariant "spark30scala213", spark30Version, scala213Version
17-
addCoreFeatureVariant "spark30scala212", spark30Version, scala212Version
18-
addCoreFeatureVariant "spark20scala211", spark24Version, scala211Version
1917

2018
all { SparkVariantPlugin.SparkVariant variant ->
2119

@@ -44,7 +42,6 @@ sparkVariants {
4442
add(variant.configuration('api'), "org.scala-lang:scala-library:${variant.scalaVersion}")
4543
add(variant.configuration('api'), "org.scala-lang:scala-reflect:${variant.scalaVersion}")
4644
add(variant.configuration('api'), "org.apache.spark:spark-core_${variant.scalaMajorVersion}:${variant.sparkVersion}") {
47-
exclude group: 'javax.servlet'
4845
exclude group: 'org.apache.hadoop'
4946
}
5047

@@ -59,29 +56,11 @@ sparkVariants {
5956
add(variant.configuration('test', 'implementation'), project(":test:shared"))
6057
add(variant.configuration('test', 'implementation'), "com.esotericsoftware:kryo:4.0.2")
6158
add(variant.configuration('test', 'implementation'), "org.apache.spark:spark-core_${variant.scalaMajorVersion}:${variant.sparkVersion}") {
62-
exclude group: 'javax.servlet'
6359
exclude group: 'org.apache.hadoop'
6460
}
6561

6662
add(variant.configuration('itest', 'implementation'), project(":test:shared"))
6763
add(variant.configuration('test', 'implementation'), "org.elasticsearch:securemock:1.2")
68-
69-
if (variant.scalaMajorVersion == '2.10') {
70-
add(variant.configuration('implementation'), "org.apache.spark:spark-unsafe_${variant.scalaMajorVersion}:${variant.sparkVersion}")
71-
add(variant.configuration('implementation'), "org.apache.avro:avro:1.7.7")
72-
add(variant.configuration('implementation'), "log4j:log4j:1.2.17")
73-
add(variant.configuration('implementation'), "com.google.code.findbugs:jsr305:2.0.1")
74-
add(variant.configuration('implementation'), "org.json4s:json4s-ast_2.10:3.2.10")
75-
add(variant.configuration('implementation'), "com.esotericsoftware.kryo:kryo:2.21")
76-
add(variant.configuration('compileOnly'), "org.apache.hadoop:hadoop-annotations:${project.ext.hadoopVersion}")
77-
add(variant.configuration('compileOnly'), "org.codehaus.jackson:jackson-core-asl:${project.ext.jacksonVersion}")
78-
add(variant.configuration('compileOnly'), "org.codehaus.jackson:jackson-mapper-asl:${project.ext.jacksonVersion}")
79-
add(variant.configuration('compileOnly'), "org.codehaus.woodstox:stax2-api:3.1.4")
80-
if (variant.sparkVersion == spark22Version) {
81-
add(variant.configuration('compileOnly'), "org.apache.spark:spark-tags_${variant.scalaMajorVersion}:${variant.sparkVersion}")
82-
}
83-
}
84-
8564
add(variant.configuration('additionalSources'), project(":elasticsearch-hadoop-mr"))
8665
add(variant.configuration('javadocSources'), project(":elasticsearch-hadoop-mr"))
8766
}
@@ -128,21 +107,6 @@ sparkVariants {
128107
}
129108
}
130109

131-
// deal with the messy conflicts out there
132-
// Ignore the scalaCompilerPlugin configurations since it is immediately resolved to configure the scala compiler tasks
133-
configurations.matching{ it.name.contains('CompilerPlugin') == false && (it.name.contains("spark30") || it.name.contains("Spark30")) == false}.all { Configuration conf ->
134-
conf.resolutionStrategy {
135-
eachDependency { details ->
136-
// change all javax.servlet artifacts to the one used by Spark otherwise these will lead to
137-
// SecurityException (signer information wrong)
138-
if (details.requested.name.contains("servlet") && !details.requested.name.contains("guice")) {
139-
details.useTarget group: "org.eclipse.jetty.orbit", name: "javax.servlet", version: "3.0.0.v201112011016"
140-
}
141-
}
142-
}
143-
conf.exclude group: "org.mortbay.jetty"
144-
}
145-
146110
if (JavaVersion.current() >= JavaVersion.VERSION_17) {
147111
tasks.withType(Test) { Test task ->
148112
if (task.getName().startsWith("test"))

spark/core/licenses/spark-core_2.12-2.4.4.jar.sha1

-1
This file was deleted.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
6f7105e792cbb650eca44890f0444720bd6a8204

0 commit comments

Comments
 (0)