-
Notifications
You must be signed in to change notification settings - Fork 232
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update task memory and disk spill metrics when buffer store spills #848
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Signed-off-by: Jason Lowe <jlowe@nvidia.com>
build |
abellina
previously approved these changes
Sep 24, 2020
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I had one comment to see if we wanted a log line from host => disk, LGTM though.
sql-plugin/src/main/scala/com/nvidia/spark/rapids/RapidsHostMemoryStore.scala
Show resolved
Hide resolved
Signed-off-by: Jason Lowe <jlowe@nvidia.com>
abellina
approved these changes
Sep 24, 2020
build |
sperlingxx
pushed a commit
to sperlingxx/spark-rapids
that referenced
this pull request
Nov 20, 2020
…VIDIA#848) * Update task memory and disk spill metrics when buffer store spills Signed-off-by: Jason Lowe <jlowe@nvidia.com> * Add log for amount of bytes spilled from host store Signed-off-by: Jason Lowe <jlowe@nvidia.com>
nartal1
pushed a commit
to nartal1/spark-rapids
that referenced
this pull request
Jun 9, 2021
…VIDIA#848) * Update task memory and disk spill metrics when buffer store spills Signed-off-by: Jason Lowe <jlowe@nvidia.com> * Add log for amount of bytes spilled from host store Signed-off-by: Jason Lowe <jlowe@nvidia.com>
nartal1
pushed a commit
to nartal1/spark-rapids
that referenced
this pull request
Jun 9, 2021
…VIDIA#848) * Update task memory and disk spill metrics when buffer store spills Signed-off-by: Jason Lowe <jlowe@nvidia.com> * Add log for amount of bytes spilled from host store Signed-off-by: Jason Lowe <jlowe@nvidia.com>
tgravescs
pushed a commit
to tgravescs/spark-rapids
that referenced
this pull request
Nov 30, 2023
…IDIA#848) Signed-off-by: spark-rapids automation <70000568+nvauto@users.noreply.github.com> Signed-off-by: spark-rapids automation <70000568+nvauto@users.noreply.github.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This connects the buffer store spill to task metrics. When a spill occurs from the device store to the host memory store, it updates the memory spilled metric for the task corresponding to the current thread that triggered the spill. Similarly when the host store spills to the disk store, it updates the disk spilled metric.
This is far from perfect, as tasks can be blocked on the memory store lock while another task is spilling and not have anything reported yet be stalled/slow because of spilling. However it at least allows a user to see that host and disk spills are occurring in their queries.