Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEA] Ability to limit total GPU memory used #488

Closed
jlowe opened this issue Jul 31, 2020 · 0 comments · Fixed by #517
Closed

[FEA] Ability to limit total GPU memory used #488

jlowe opened this issue Jul 31, 2020 · 0 comments · Fixed by #517
Assignees
Labels
feature request New feature or request P1 Nice to have for release

Comments

@jlowe
Copy link
Member

jlowe commented Jul 31, 2020

Is your feature request related to a problem? Please describe.
Sometimes the plugin needs to share the GPU with another process, and the plugin's current behavior of aggressively allocating most of the GPU memory for the RMM pool is hostile to that setup. Even if one manually configures the initial RMM pool size to be lower, it will grow without bound if the plugin requests that much memory. It would be nice to be able to put a hard limit on the RMM memory size.

Describe the solution you'd like
A plugin config that can be used to set a limit on the total size of the RMM memory pool.

Describe alternatives you've considered
Being able to share the same RMM pool between the processes would be ideal, then we wouldn't have to hard-partition the GPU like this. However even if that feature existed, we would still need some guarantee the plugin wouldn't end up consuming all of the shared RMM pool just as the other process needs it.

@jlowe jlowe added feature request New feature or request ? - Needs Triage Need team to review and classify P1 Nice to have for release labels Jul 31, 2020
@jlowe jlowe removed the ? - Needs Triage Need team to review and classify label Jul 31, 2020
@jlowe jlowe self-assigned this Aug 4, 2020
@jlowe jlowe added this to the Aug 3 - Aug 14 milestone Aug 4, 2020
pxLi pushed a commit to pxLi/spark-rapids that referenced this issue May 12, 2022
tgravescs pushed a commit to tgravescs/spark-rapids that referenced this issue Nov 30, 2023
[auto-merge] bot-auto-merge-branch-22.08 to branch-22.10 [skip ci] [bot]
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request P1 Nice to have for release
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant