Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Updated GCP Dataproc Mortgage-ETL-GPU.ipynb #991

Merged
merged 3 commits into from
Oct 26, 2020

Conversation

aroraakshit
Copy link
Contributor

Description of changes:

  • from pyspark import SparkConf
  • if sc: sc.stop()
  • Added instructions to download and organize the dataset in a GCP bucket.

@sameerz sameerz added the documentation Improvements or additions to documentation label Oct 21, 2020
@sameerz sameerz added this to the Oct 12 - Oct 23 milestone Oct 21, 2020
@jlowe
Copy link
Member

jlowe commented Oct 21, 2020

Thanks for the contribution, @aroraakshit! As reported by the CI checks, you will need to sign-off on your contribution per the instructions in the contribution guide. Easiest would be to perform a git commit --amend -s and then force-pushing your branch back up to github.

@sameerz sameerz self-requested a review October 22, 2020 00:08
Signed-off-by: Akshit Arora <akshita@nvidia.com>
Copy link
Contributor Author

@aroraakshit aroraakshit left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

if sc: sc.stop() will fail if 'sc' variable does not exist. So changed to if 'sc' in globals().

@jlowe
Copy link
Member

jlowe commented Oct 24, 2020

build

1 similar comment
@jlowe
Copy link
Member

jlowe commented Oct 24, 2020

build

Copy link
Collaborator

@revans2 revans2 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The pre-commit build is failing because of the 3.1 changes to parquet. Either this needs to be upmerged or we can ignore the build failure and merge it in because this is not executed by the pre-commit build at all.

@jlowe
Copy link
Member

jlowe commented Oct 26, 2020

Merging since the CI failure is unrelated to the notebook change.

@jlowe jlowe merged commit 1b58b68 into NVIDIA:branch-0.3 Oct 26, 2020
sperlingxx pushed a commit to sperlingxx/spark-rapids that referenced this pull request Nov 20, 2020
* updated notebook

* variable check updated

* signed

Signed-off-by: Akshit Arora <akshita@nvidia.com>
nartal1 pushed a commit to nartal1/spark-rapids that referenced this pull request Jun 9, 2021
* updated notebook

* variable check updated

* signed

Signed-off-by: Akshit Arora <akshita@nvidia.com>
nartal1 pushed a commit to nartal1/spark-rapids that referenced this pull request Jun 9, 2021
* updated notebook

* variable check updated

* signed

Signed-off-by: Akshit Arora <akshita@nvidia.com>
tgravescs pushed a commit to tgravescs/spark-rapids that referenced this pull request Nov 30, 2023
…IDIA#991)

Signed-off-by: spark-rapids automation <70000568+nvauto@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants