Skip to content

Commit 61651e0

Browse files
authored
[SW-2138] Apply spotless formatting to gradle files (#2007)
1 parent b7cc011 commit 61651e0

File tree

25 files changed

+1096
-1090
lines changed

25 files changed

+1096
-1090
lines changed

apps/streaming/build.gradle

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
description = "Sparkling Water Pipelining"
22

3-
// Support local launch of application
3+
// Support local launch of application
44
apply plugin: 'application'
55
mainClassName = "ai.h2o.demo.PipelineDemo"
66
// Disable distribution tasks
@@ -18,4 +18,3 @@ dependencies {
1818
compileOnly("org.apache.spark:spark-streaming_${scalaBaseVersion}:${sparkVersion}")
1919
compileOnly("org.scala-lang:scala-library:${scalaVersion}")
2020
}
21-

assembly-extensions/build.gradle

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -5,29 +5,29 @@ apply plugin: 'java-library'
55
description = "Sparkling Water Extensions Assembly"
66

77
dependencies {
8-
api("org.scala-lang:scala-library:${scalaVersion}")
9-
api(project(":sparkling-water-extensions"))
10-
api(project(":sparkling-water-utils")) { transitive = false }
8+
api("org.scala-lang:scala-library:${scalaVersion}")
9+
api(project(":sparkling-water-extensions"))
10+
api(project(":sparkling-water-utils")) { transitive = false }
1111
}
1212

1313
jar {
14-
enabled = false // we do not need to generate jar file in this case
14+
enabled = false // we do not need to generate jar file in this case
1515
}
1616

1717
shadowJar {
18-
mergeServiceFiles()
18+
mergeServiceFiles()
1919

20-
archiveBaseName = "${archiveBaseName.get()}_${scalaBaseVersion}"
20+
archiveBaseName = "${archiveBaseName.get()}_${scalaBaseVersion}"
2121
}
2222

2323
artifacts {
24-
api shadowJar
24+
api shadowJar
2525
}
2626

2727
task embeddedAssemblyJar(type: Copy, dependsOn: shadowJar) {
28-
from file("$buildDir/libs")
29-
into file("$buildDir/embedded")
30-
rename { _ -> 'assembly-extensions.jar.embedded' }
28+
from file("$buildDir/libs")
29+
into file("$buildDir/embedded")
30+
rename { _ -> 'assembly-extensions.jar.embedded' }
3131
}
3232

3333
build.dependsOn embeddedAssemblyJar

assembly/build.gradle

Lines changed: 42 additions & 42 deletions
Original file line numberDiff line numberDiff line change
@@ -9,70 +9,70 @@ description = "Sparkling Water Assembly"
99

1010

1111
dependencies {
12-
api(project(":sparkling-water-ml"))
13-
api(project(":sparkling-water-repl"))
14-
api(project(":sparkling-water-core"))
15-
api(project(":sparkling-water-extensions"))
12+
api(project(":sparkling-water-ml"))
13+
api(project(":sparkling-water-repl"))
14+
api(project(":sparkling-water-core"))
15+
api(project(":sparkling-water-extensions"))
1616
}
1717

1818
//
1919
// Contains runtime configuration for Sparkling Water fat jar. This configuration transitively removes dependencies provided by
2020
// Apache Spark or the Hadoop ecosystem.
2121
//
2222
configurations {
23-
shadowApi {
24-
extendsFrom api
25-
exclude group: 'org.apache.spark'
26-
exclude group: 'org.apache.hadoop'
27-
exclude group: 'org.scala-lang'
28-
exclude group: "javax.servlet", module: "servlet-api"
29-
exclude group: 'org.apache.commons', module: 'commons-math3' // a dependency of org.apache.spark:spark-core_2.11
30-
exclude group: 'org.codehaus.jackson', module: 'jackson-core-asl'
31-
// a dependency of org.apache.spark:spark-core_2.11
32-
exclude group: 'commons-codec', module: 'commons-codec' // a dependency of org.apache.spark:spark-sql_2.11
33-
exclude group: 'commons-lang', module: 'commons-lang' // a dependency of org.apache.spark:spark-core_2.11
34-
exclude group: 'commons-io', module: 'commons-io' // a dependency of org.apache.spark:spark-core_2.11
35-
exclude group: 'commons-logging', module: 'commons-logging' // a dependency of org.apache.hadoop:hadoop-auth
36-
exclude group: 'log4j', module: 'log4j' // a dependency of org.apache.hadoop:hadoop-auth
37-
exclude group: 'com.google.protobuf' // a dependency of org.apache.hadoop:hadoop-common
38-
exclude group: 'com.fasterxml.jackson.core', module: 'jackson-core'
39-
// a dependency of org.apache.spark:spark-sql_2.11
40-
exclude group: 'org.apache.httpcomponents' // a dependency of org.apache.hadoop:hadoop-auth
41-
exclude group: 'com.github.rwl', module: 'jtransforms' // a dependency of org.apache.spark:spark-mllib
42-
exclude group: 'com.google.code.findbugs', module: 'jsr305' // a dependency of org.apache.hadoop:hadoop-common
43-
exclude group: 'javax.xml.bind', module: 'jaxb-api' // a dependency of org.apache.hadoop:hadoop-yarn-common
44-
exclude group: 'net.sourceforge.f2j', module: 'arpack_combined_all'
45-
// a dependency of org.apache.spark:spark-graphx_2.11
46-
}
23+
shadowApi {
24+
extendsFrom api
25+
exclude group: 'org.apache.spark'
26+
exclude group: 'org.apache.hadoop'
27+
exclude group: 'org.scala-lang'
28+
exclude group: "javax.servlet", module: "servlet-api"
29+
exclude group: 'org.apache.commons', module: 'commons-math3' // a dependency of org.apache.spark:spark-core_2.11
30+
exclude group: 'org.codehaus.jackson', module: 'jackson-core-asl'
31+
// a dependency of org.apache.spark:spark-core_2.11
32+
exclude group: 'commons-codec', module: 'commons-codec' // a dependency of org.apache.spark:spark-sql_2.11
33+
exclude group: 'commons-lang', module: 'commons-lang' // a dependency of org.apache.spark:spark-core_2.11
34+
exclude group: 'commons-io', module: 'commons-io' // a dependency of org.apache.spark:spark-core_2.11
35+
exclude group: 'commons-logging', module: 'commons-logging' // a dependency of org.apache.hadoop:hadoop-auth
36+
exclude group: 'log4j', module: 'log4j' // a dependency of org.apache.hadoop:hadoop-auth
37+
exclude group: 'com.google.protobuf' // a dependency of org.apache.hadoop:hadoop-common
38+
exclude group: 'com.fasterxml.jackson.core', module: 'jackson-core'
39+
// a dependency of org.apache.spark:spark-sql_2.11
40+
exclude group: 'org.apache.httpcomponents' // a dependency of org.apache.hadoop:hadoop-auth
41+
exclude group: 'com.github.rwl', module: 'jtransforms' // a dependency of org.apache.spark:spark-mllib
42+
exclude group: 'com.google.code.findbugs', module: 'jsr305' // a dependency of org.apache.hadoop:hadoop-common
43+
exclude group: 'javax.xml.bind', module: 'jaxb-api' // a dependency of org.apache.hadoop:hadoop-yarn-common
44+
exclude group: 'net.sourceforge.f2j', module: 'arpack_combined_all'
45+
// a dependency of org.apache.spark:spark-graphx_2.11
46+
}
4747
}
4848

4949
jar {
50-
enabled = false // we do not need to generate jar file in this case
51-
archiveBaseName = "${jar.archiveBaseName.get()}_${scalaBaseVersion}"
50+
enabled = false // we do not need to generate jar file in this case
51+
archiveBaseName = "${jar.archiveBaseName.get()}_${scalaBaseVersion}"
5252
}
5353

5454
shadowJar {
55-
configurations = [project.configurations.shadowApi]
56-
mergeServiceFiles()
55+
configurations = [project.configurations.shadowApi]
56+
mergeServiceFiles()
5757

58-
relocate 'javassist', 'ai.h2o.javassist'
59-
relocate 'com.google.common', 'ai.h2o.com.google.common'
60-
relocate 'org.eclipse.jetty', 'ai.h2o.org.eclipse.jetty'
61-
relocate 'org.eclipse.jetty.orbit', 'ai.h2o.org.eclipse.jetty.orbit'
58+
relocate 'javassist', 'ai.h2o.javassist'
59+
relocate 'com.google.common', 'ai.h2o.com.google.common'
60+
relocate 'org.eclipse.jetty', 'ai.h2o.org.eclipse.jetty'
61+
relocate 'org.eclipse.jetty.orbit', 'ai.h2o.org.eclipse.jetty.orbit'
6262

63-
from "$project.buildDir/reports/" include '**/*'
64-
exclude 'www/flow/packs/test-*/**'
63+
from "$project.buildDir/reports/" include '**/*'
64+
exclude 'www/flow/packs/test-*/**'
6565

66-
archiveBaseName = "${archiveBaseName.get()}_${scalaBaseVersion}"
66+
archiveBaseName = "${archiveBaseName.get()}_${scalaBaseVersion}"
6767
}
6868

6969
licenseReport {
70-
allowedLicensesFile = new File("$projectDir/allowed-licenses.json")
71-
filters = new LicenseBundleNormalizer(bundlePath: "$projectDir/license-normalizer-bundle.json")
70+
allowedLicensesFile = new File("$projectDir/allowed-licenses.json")
71+
filters = new LicenseBundleNormalizer(bundlePath: "$projectDir/license-normalizer-bundle.json")
7272
}
7373

7474
artifacts {
75-
api shadowJar
75+
api shadowJar
7676
}
7777

7878
shadowJar.dependsOn generateLicenseReport

benchmarks/build.gradle

Lines changed: 96 additions & 95 deletions
Original file line numberDiff line numberDiff line change
@@ -4,131 +4,132 @@ apply from: "$rootDir/gradle/utils.gradle"
44
apply plugin: 'java-library'
55

66
configurations {
7-
sparklingWaterAssemblyJar
7+
sparklingWaterAssemblyJar
88
}
99

1010
dependencies {
11-
sparklingWaterAssemblyJar(project(path: ':sparkling-water-assembly', configuration: 'shadow'))
11+
sparklingWaterAssemblyJar(project(path: ':sparkling-water-assembly', configuration: 'shadow'))
1212

13-
api(project(":sparkling-water-ml"))
14-
api(project(":sparkling-water-core"))
13+
api(project(":sparkling-water-ml"))
14+
api(project(":sparkling-water-core"))
1515

16-
compileOnly("org.apache.spark:spark-sql_${scalaBaseVersion}:${sparkVersion}")
17-
compileOnly("org.apache.spark:spark-mllib_${scalaBaseVersion}:${sparkVersion}")
16+
compileOnly("org.apache.spark:spark-sql_${scalaBaseVersion}:${sparkVersion}")
17+
compileOnly("org.apache.spark:spark-mllib_${scalaBaseVersion}:${sparkVersion}")
1818
}
1919

2020
task cleanTerraform(type: Delete) {
21-
delete "build/terraform"
21+
delete "build/terraform"
2222
}
2323

2424
task copyTerraform(dependsOn: cleanTerraform) {
25-
doLast {
26-
copy {
27-
from 'src/main/terraform'
28-
include "**/*.tf"
29-
into "build/terraform"
30-
}
31-
copy {
32-
from '../templates/src/terraform/aws/modules'
33-
include "emr_security/*.tf"
34-
into "build/terraform/aws/modules"
35-
}
25+
doLast {
26+
copy {
27+
from 'src/main/terraform'
28+
include "**/*.tf"
29+
into "build/terraform"
3630
}
31+
copy {
32+
from '../templates/src/terraform/aws/modules'
33+
include "emr_security/*.tf"
34+
into "build/terraform/aws/modules"
35+
}
36+
}
3737
}
3838

3939
task substituteTerraform(dependsOn: copyTerraform) {
40-
doLast {
41-
def tfBaseDir = "${project.buildDir.toString()}/terraform/aws/"
42-
def tfScripts = [
43-
"${tfBaseDir}/variables.tf",
44-
"${tfBaseDir}/modules/emr_benchmarks_deployment/variables.tf",
45-
"${tfBaseDir}/modules/emr_benchmarks_deployment/main.tf"]
46-
tfScripts.each { path ->
47-
def contents = file(path).getText('UTF-8')
48-
contents = contents
49-
.replaceAll("SUBST_PACKAGE_FILE", "${configurations.sparklingWaterAssemblyJar.singleFile}")
50-
.replaceAll("SUBST_BENCHMARKS_FILE", "$buildDir/libs/sparkling-water-benchmarks_$scalaBaseVersion-${version}.jar")
51-
.replaceAll("SUBST_H2O_VERSION_NAME", h2oMajorName)
52-
.replaceAll("SUBST_H2O_VERSION", h2oVersion)
53-
.replaceAll("SUBST_H2O_BUILD", h2oBuild)
54-
.replaceAll("SUBST_SW_VERSION", version.toString())
55-
.replaceAll("SUBST_SCALA_VERSION", scalaBaseVersion)
56-
.replaceAll("SUBST_EMR_VERSION", supportedEmrVersion)
57-
58-
file(path).write(contents, 'UTF-8')
59-
}
40+
doLast {
41+
def tfBaseDir = "${project.buildDir.toString()}/terraform/aws/"
42+
def tfScripts = [
43+
"${tfBaseDir}/variables.tf",
44+
"${tfBaseDir}/modules/emr_benchmarks_deployment/variables.tf",
45+
"${tfBaseDir}/modules/emr_benchmarks_deployment/main.tf"
46+
]
47+
tfScripts.each { path ->
48+
def contents = file(path).getText('UTF-8')
49+
contents = contents
50+
.replaceAll("SUBST_PACKAGE_FILE", "${configurations.sparklingWaterAssemblyJar.singleFile}")
51+
.replaceAll("SUBST_BENCHMARKS_FILE", "$buildDir/libs/sparkling-water-benchmarks_$scalaBaseVersion-${version}.jar")
52+
.replaceAll("SUBST_H2O_VERSION_NAME", h2oMajorName)
53+
.replaceAll("SUBST_H2O_VERSION", h2oVersion)
54+
.replaceAll("SUBST_H2O_BUILD", h2oBuild)
55+
.replaceAll("SUBST_SW_VERSION", version.toString())
56+
.replaceAll("SUBST_SCALA_VERSION", scalaBaseVersion)
57+
.replaceAll("SUBST_EMR_VERSION", supportedEmrVersion)
58+
59+
file(path).write(contents, 'UTF-8')
6060
}
61+
}
6162
}
6263

6364
task cleanOutput(type: Delete) {
64-
delete "output"
65+
delete "output"
6566
}
6667

6768
task runBenchmarks(dependsOn: [":sparkling-water-assembly:shadowJar", substituteTerraform, cleanOutput]) {
68-
doLast {
69-
exec {
70-
def accessKey = project.property("aws_access_key")
71-
def secretKey = project.property("aws_secret_key")
72-
def publicKey = project.property("aws_ssh_public_key")
73-
74-
environment("aws_access_key", accessKey)
75-
environment("aws_secret_key", secretKey)
76-
environment("aws_ssh_public_key", publicKey)
77-
environment("datasets", "datasets.json")
78-
79-
commandLine "./run_benchmarks.sh"
80-
}
69+
doLast {
70+
exec {
71+
def accessKey = project.property("aws_access_key")
72+
def secretKey = project.property("aws_secret_key")
73+
def publicKey = project.property("aws_ssh_public_key")
74+
75+
environment("aws_access_key", accessKey)
76+
environment("aws_secret_key", secretKey)
77+
environment("aws_ssh_public_key", publicKey)
78+
environment("datasets", "datasets.json")
79+
80+
commandLine "./run_benchmarks.sh"
8181
}
82+
}
8283
}
8384

8485
task runBigDataSparkToH2OConversionBenchmarks(dependsOn: [":sparkling-water-assembly:shadowJar", substituteTerraform, cleanOutput]) {
85-
doLast {
86-
exec {
87-
def accessKey = project.property("aws_access_key")
88-
def secretKey = project.property("aws_secret_key")
89-
def publicKey = project.property("aws_ssh_public_key")
90-
91-
environment("aws_access_key", accessKey)
92-
environment("aws_secret_key", secretKey)
93-
environment("aws_ssh_public_key", publicKey)
94-
environment("aws_instance_type", "m5.4xlarge")
95-
environment("aws_core_instance_count", "10")
96-
environment("datasets", "bigDatasets.json")
97-
environment("other_arguments", "-b DataFrameToH2OFrameConversionBenchmark")
98-
environment("driver_memory_gb", "8")
99-
environment("executor_memory_gb", "32")
100-
environment("run_yarn_internal", "false")
101-
environment("run_yarn_external", "true")
102-
environment("run_local_internal", "false")
103-
104-
commandLine "./run_benchmarks.sh"
105-
}
86+
doLast {
87+
exec {
88+
def accessKey = project.property("aws_access_key")
89+
def secretKey = project.property("aws_secret_key")
90+
def publicKey = project.property("aws_ssh_public_key")
91+
92+
environment("aws_access_key", accessKey)
93+
environment("aws_secret_key", secretKey)
94+
environment("aws_ssh_public_key", publicKey)
95+
environment("aws_instance_type", "m5.4xlarge")
96+
environment("aws_core_instance_count", "10")
97+
environment("datasets", "bigDatasets.json")
98+
environment("other_arguments", "-b DataFrameToH2OFrameConversionBenchmark")
99+
environment("driver_memory_gb", "8")
100+
environment("executor_memory_gb", "32")
101+
environment("run_yarn_internal", "false")
102+
environment("run_yarn_external", "true")
103+
environment("run_local_internal", "false")
104+
105+
commandLine "./run_benchmarks.sh"
106106
}
107+
}
107108
}
108109

109110
task runBigDataH2OtoSparkConversionBenchmarks(dependsOn: [":sparkling-water-assembly:shadowJar", substituteTerraform, cleanOutput]) {
110-
doLast {
111-
exec {
112-
def accessKey = project.property("aws_access_key")
113-
def secretKey = project.property("aws_secret_key")
114-
def publicKey = project.property("aws_ssh_public_key")
115-
116-
environment("aws_access_key", accessKey)
117-
environment("aws_secret_key", secretKey)
118-
environment("aws_ssh_public_key", publicKey)
119-
environment("aws_instance_type", "m5.4xlarge")
120-
environment("aws_core_instance_count", "10")
121-
environment("datasets", "bigDatasets.json")
122-
environment("other_arguments", "-b H2OFrameToDataFrameConversionBenchmark")
123-
environment("driver_memory_gb", "8")
124-
environment("executor_memory_gb", "32")
125-
environment("run_yarn_internal", "false")
126-
environment("run_yarn_external", "true")
127-
environment("run_local_internal", "false")
128-
129-
commandLine "./run_benchmarks.sh"
130-
}
111+
doLast {
112+
exec {
113+
def accessKey = project.property("aws_access_key")
114+
def secretKey = project.property("aws_secret_key")
115+
def publicKey = project.property("aws_ssh_public_key")
116+
117+
environment("aws_access_key", accessKey)
118+
environment("aws_secret_key", secretKey)
119+
environment("aws_ssh_public_key", publicKey)
120+
environment("aws_instance_type", "m5.4xlarge")
121+
environment("aws_core_instance_count", "10")
122+
environment("datasets", "bigDatasets.json")
123+
environment("other_arguments", "-b H2OFrameToDataFrameConversionBenchmark")
124+
environment("driver_memory_gb", "8")
125+
environment("executor_memory_gb", "32")
126+
environment("run_yarn_internal", "false")
127+
environment("run_yarn_external", "true")
128+
environment("run_local_internal", "false")
129+
130+
commandLine "./run_benchmarks.sh"
131131
}
132+
}
132133
}
133134

134-
substituteTerraform.dependsOn build
135+
substituteTerraform.dependsOn build

0 commit comments

Comments
 (0)