Skip to content

Commit

Permalink
Update to 0.2.0-SNAPSHOT and cudf-0.15-SNAPSHOT (NVIDIA#252)
Browse files Browse the repository at this point in the history
* Update to 0.2.0-SNAPSHOT and cudf-0.15-SNAPSHOT

* Update a few missed places for cudf-0.15-SNAPSHOT
  • Loading branch information
jlowe authored Jun 22, 2020
1 parent d7b488e commit 5a7dfed
Show file tree
Hide file tree
Showing 16 changed files with 32 additions and 32 deletions.
4 changes: 2 additions & 2 deletions api_validation/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,10 +22,10 @@
<parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-parent</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>
</parent>
<artifactId>rapids-4-spark-api-validation</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>

<dependencies>
<dependency>
Expand Down
4 changes: 2 additions & 2 deletions dist/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,13 +22,13 @@
<parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-parent</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>
</parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark_2.12</artifactId>
<name>RAPIDS Accelerator for Apache Spark Distribution</name>
<description>Creates the distribution package of the RAPIDS plugin for Apache Spark</description>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>

<dependencies>
<dependency>
Expand Down
2 changes: 1 addition & 1 deletion docs/configs.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ The following is the list of options that `rapids-plugin-4-spark` supports.
On startup use: `--conf [conf key]=[conf value]`. For example:

```
${SPARK_HOME}/bin/spark --jars 'rapids-4-spark_2.12-0.1-SNAPSHOT.jar,cudf-0.14-SNAPSHOT-cuda10.jar' \
${SPARK_HOME}/bin/spark --jars 'rapids-4-spark_2.12-0.2.0-SNAPSHOT.jar,cudf-0.15-SNAPSHOT-cuda10-1.jar' \
--conf spark.plugins=com.nvidia.spark.SQLPlugin \
--conf spark.rapids.sql.incompatibleOps.enabled=true
```
Expand Down
6 changes: 3 additions & 3 deletions docs/getting-started.md
Original file line number Diff line number Diff line change
Expand Up @@ -108,15 +108,15 @@ CUDA and will not run on other versions. The jars use a maven classifier to keep
- CUDA 10.2 => classifier cuda10-2

For example, here is a sample version of the jars and cudf with CUDA 10.1 support:
- cudf-0.14-cuda10-1.jar
- cudf-0.15-SNAPSHOT-cuda10-1.jar
- rapids-4-spark_2.12-0.1.0.jar

For simplicity export the location to these jars. This example assumes the sample jars above have
been placed in the `/opt/sparkRapidsPlugin` directory:
```shell
export SPARK_RAPIDS_DIR=/opt/sparkRapidsPlugin
export SPARK_CUDF_JAR=${SPARK_RAPIDS_DIR}/cudf-0.14-cuda10-1.jar
export SPARK_RAPIDS_PLUGIN_JAR=${SPARK_RAPIDS_DIR}/rapids-4-spark_2.12-0.1.0.jar
export SPARK_CUDF_JAR=${SPARK_RAPIDS_DIR}/cudf-0.15-SNAPSHOT-cuda10-1.jar
export SPARK_RAPIDS_PLUGIN_JAR=${SPARK_RAPIDS_DIR}/rapids-4-spark_2.12-0.2.0-SNAPSHOT.jar
```

## Install the GPU Discovery Script
Expand Down
4 changes: 2 additions & 2 deletions docs/testing.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@ we typically run with the default options and only increase the scale factor dep
dbgen -b dists.dss -s 10
```

You can include the test jar `rapids-4-spark-integration-tests_2.12-0.1-SNAPSHOT.jar` with the
You can include the test jar `rapids-4-spark-integration-tests_2.12-0.2.0-SNAPSHOT.jar` with the
Spark --jars option to get the TPCH tests. To setup for the queries you can run
`TpchLikeSpark.setupAllCSV` for CSV formatted data or `TpchLikeSpark.setupAllParquet`
for parquet formatted data. Both of those take the Spark session, and a path to the dbgen
Expand Down Expand Up @@ -77,7 +77,7 @@ individually, so you don't risk running unit tests along with the integration te
http://www.scalatest.org/user_guide/using_the_scalatest_shell

```shell
spark-shell --jars rapids-4-spark-tests_2.12-0.1-SNAPSHOT-tests.jar,rapids-4-spark-integration-tests_2.12-0.1-SNAPSHOT-tests.jar,scalatest_2.12-3.0.5.jar,scalactic_2.12-3.0.5.jar
spark-shell --jars rapids-4-spark-tests_2.12-0.2.0-SNAPSHOT-tests.jar,rapids-4-spark-integration-tests_2.12-0.2.0-SNAPSHOT-tests.jar,scalatest_2.12-3.0.5.jar,scalactic_2.12-3.0.5.jar
```

First you import the `scalatest_shell` and tell the tests where they can find the test files you
Expand Down
4 changes: 2 additions & 2 deletions integration_tests/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ Most clusters probably will not have the RAPIDS plugin installed in the cluster
If just want to verify the SQL replacement is working you will need to add the `rapids-4-spark` and `cudf` jars to your `spark-submit` command.

```
$SPARK_HOME/bin/spark-submit --jars "rapids-4-spark_2.12-0.1-SNAPSHOT.jar,cudf-0.14.jar" ./runtests.py
$SPARK_HOME/bin/spark-submit --jars "rapids-4-spark_2.12-0.2.0-SNAPSHOT.jar,cudf-0.15-SNAPSHOT.jar" ./runtests.py
```

You don't have to enable the plugin for this to work, the test framework will do that for you.
Expand Down Expand Up @@ -70,7 +70,7 @@ The TPCxBB, TPCH, and Mortgage tests in this framework can be enabled by providi
As an example, here is the `spark-submit` command with the TPCxBB parameters:

```
$SPARK_HOME/bin/spark-submit --jars "rapids-4-spark_2.12-0.1-SNAPSHOT.jar,cudf-0.14.jar,rapids-4-spark-tests_2.12-0.1-SNAPSHOT.jar" ./runtests.py --tpcxbb_format="csv" --tpcxbb_path="/path/to/tpcxbb/csv"
$SPARK_HOME/bin/spark-submit --jars "rapids-4-spark_2.12-0.2.0-SNAPSHOT.jar,cudf-0.15-SNAPSHOT.jar,rapids-4-spark-tests_2.12-0.2.0-SNAPSHOT.jar" ./runtests.py --tpcxbb_format="csv" --tpcxbb_path="/path/to/tpcxbb/csv"
```

## Writing tests
Expand Down
4 changes: 2 additions & 2 deletions integration_tests/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,11 +22,11 @@
<parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-parent</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>
</parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-integration-tests_2.12</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>

<dependencies>
<dependency>
Expand Down
8 changes: 4 additions & 4 deletions jenkins/Jenkinsfile.integration
Original file line number Diff line number Diff line change
Expand Up @@ -32,17 +32,17 @@ pipeline {
}

parameters {
string(name: 'CUDF_VER', defaultValue: '0.14',
string(name: 'CUDF_VER', defaultValue: '0.15-SNAPSHOT',
description: '-Dcudf.version= \n\n Default for cudf version')
string(name: 'CUDA_CLASSIFIER', defaultValue: '',
description: '-Dclassifier=\n\n cuda10-1, cuda10-2, EMPTY as cuda10-1')
string(name: 'PROJECT_VER', defaultValue: '0.1-SNAPSHOT',
description: 'Default project version 0.1-SNAPSHOT')
string(name: 'PROJECT_VER', defaultValue: '0.2.0-SNAPSHOT',
description: 'Default project version 0.2.0-SNAPSHOT')
string(name: 'SPARK_VER', defaultValue: '3.0.0',
description: 'Default spark version 3.0.0')
string(name: 'SERVER_URL', defaultValue: 'https://urm.nvidia.com:443/artifactory/sw-spark-maven',
description: 'Default maven repo URL where to download Spark3.0 tar file.')
string(name: 'REF', defaultValue: 'branch-0.1', description: 'Commit to build')
string(name: 'REF', defaultValue: 'branch-0.2', description: 'Commit to build')
}

environment {
Expand Down
4 changes: 2 additions & 2 deletions jenkins/Jenkinsfile.nightly
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ pipeline {
-v ${HOME}/.zinc:${HOME}/.zinc:rw \
-v /etc/passwd:/etc/passwd -v /etc/group:/etc/group") {
sh "mvn -U -B clean deploy $MVN_URM_MIRROR"
sh "jenkins/printJarVersion.sh 'CUDFVersion' '${HOME}/.m2/repository/ai/rapids/cudf/0.14' 'cudf-0.14' '-cuda10-1.jar'"
sh "jenkins/printJarVersion.sh 'CUDFVersion' '${HOME}/.m2/repository/ai/rapids/cudf/0.15-SNAPSHOT' 'cudf-0.15-SNAPSHOT' '-cuda10-1.jar'"
sh "jenkins/printJarVersion.sh 'SPARKVersion' '${HOME}/.m2/repository/org/apache/spark/spark-core_2.12/3.0.0' 'spark-core_2.12-3.0.0-' '.jar'"
}
}
Expand All @@ -78,7 +78,7 @@ pipeline {
build(job: 'spark/rapids_integration-0.1-github',
propagate: false,
parameters: [string(name: 'REF', value: 'branch-0.1'),
string(name: 'CUDF_VER', value: '0.14'),
string(name: 'CUDF_VER', value: '0.15-SNAPSHOT'),
booleanParam(name: 'BUILD_CENTOS7', value: false),])

slack("#rapidsai-spark-cicd", "Success", color: "#33CC33")
Expand Down
2 changes: 1 addition & 1 deletion jenkins/databricks/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,7 @@ mvn install:install-file \
mvn -Pdatabricks clean verify -DskipTests

# copy so we pick up new built jar
sudo cp dist/target/rapids-4-spark_2.12-*-SNAPSHOT.jar /databricks/jars/rapids-4-spark_2.12-0.1-SNAPSHOT-ci.jar
sudo cp dist/target/rapids-4-spark_2.12-*-SNAPSHOT.jar /databricks/jars/rapids-4-spark_2.12-0.2.0-SNAPSHOT-ci.jar

# tests
export PATH=/databricks/conda/envs/databricks-ml-gpu/bin:/databricks/conda/condabin:$PATH
Expand Down
4 changes: 2 additions & 2 deletions jenkins/spark-tests.sh
Original file line number Diff line number Diff line change
Expand Up @@ -17,11 +17,11 @@

set -ex
if [ "$CUDF_VER"x == x ];then
CUDF_VER="0.14"
CUDF_VER="0.15-SNAPSHOT"
fi

if [ "$PROJECT_VER"x == x ];then
PROJECT_VER="0.1-SNAPSHOT"
PROJECT_VER="0.2.0-SNAPSHOT"
fi

if [ "$SPARK_VER"x == x ];then
Expand Down
4 changes: 2 additions & 2 deletions pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
<artifactId>rapids-4-spark-parent</artifactId>
<name>RAPIDS Accelerator for Apache Spark Root Project</name>
<description>The root project of the RAPIDS Accelerator for Apache Spark</description>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>
<packaging>pom</packaging>

<url>https://github.com/NVIDIA</url>
Expand Down Expand Up @@ -132,7 +132,7 @@
<maven.compiler.target>1.8</maven.compiler.target>
<spark.version>3.0.0</spark.version>
<cuda.version>cuda10-1</cuda.version>
<cudf.version>0.14</cudf.version>
<cudf.version>0.15-SNAPSHOT</cudf.version>
<scala.binary.version>2.12</scala.binary.version>
<scala.version>2.12.8</scala.version>
<orc.version>1.5.8</orc.version>
Expand Down
4 changes: 2 additions & 2 deletions shuffle-plugin/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,14 +22,14 @@
<parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-parent</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>
</parent>

<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-shuffle_2.12</artifactId>
<name>RAPIDS Accelerator for Apache Spark Shuffle Plugin</name>
<description>Accelerated shuffle plugin for the RAPIDS plugin for Apache Spark</description>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>

<dependencies>
<dependency>
Expand Down
4 changes: 2 additions & 2 deletions sql-plugin/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,13 +22,13 @@
<parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-parent</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>
</parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-sql_2.12</artifactId>
<name>RAPIDS Accelerator for Apache Spark SQL Plugin</name>
<description>The RAPIDS SQL plugin for Apache Spark</description>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>

<profiles>
<profile>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -612,7 +612,7 @@ object RapidsConf {
|On startup use: `--conf [conf key]=[conf value]`. For example:
|
|```
|${SPARK_HOME}/bin/spark --jars 'rapids-4-spark_2.12-0.1-SNAPSHOT.jar,cudf-0.14-SNAPSHOT-cuda10.jar' \
|${SPARK_HOME}/bin/spark --jars 'rapids-4-spark_2.12-0.2.0-SNAPSHOT.jar,cudf-0.15-SNAPSHOT-cuda10-1.jar' \
|--conf spark.plugins=com.nvidia.spark.SQLPlugin \
|--conf spark.rapids.sql.incompatibleOps.enabled=true
|```
Expand Down
4 changes: 2 additions & 2 deletions tests/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,13 +22,13 @@
<parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-parent</artifactId>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>
</parent>
<groupId>com.nvidia</groupId>
<artifactId>rapids-4-spark-tests_2.12</artifactId>
<name>RAPIDS Accelerator for Apache Spark Tests</name>
<description>RAPIDS plugin for Apache Spark integration tests</description>
<version>0.1-SNAPSHOT</version>
<version>0.2.0-SNAPSHOT</version>

<dependencies>
<dependency>
Expand Down

0 comments on commit 5a7dfed

Please sign in to comment.