Skip to content

Commit

Permalink
Use --ipc=host in docker run for distributed inference (vllm-projec…
Browse files Browse the repository at this point in the history
  • Loading branch information
WoosukKwon authored Sep 22, 2023
1 parent f98b745 commit 7d7e3b7
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion docs/source/getting_started/installation.rst
Original file line number Diff line number Diff line change
Expand Up @@ -46,4 +46,5 @@ You can also build and install vLLM from source:
.. code-block:: console
$ # Pull the Docker image with CUDA 11.8.
$ docker run --gpus all -it --rm --shm-size=8g nvcr.io/nvidia/pytorch:22.12-py3
$ # Use `--ipc=host` to make sure the shared memory is large enough.
$ docker run --gpus all -it --rm --ipc=host nvcr.io/nvidia/pytorch:22.12-py3

0 comments on commit 7d7e3b7

Please sign in to comment.