浏览代码

change gpu parallel size docs

Justin Lee 3 月之前
父节点
当前提交
03f2b8eddd
共有 1 个文件被更改,包括 1 次插入1 次删除
  1. 1 1
      end-to-end-use-cases/prompt-migration/README.md

+ 1 - 1
end-to-end-use-cases/prompt-migration/README.md

@@ -30,7 +30,7 @@ The **Prompt Migration** toolkit helps you assess and adapt prompts across diffe
    ```
    ```
    To serve a large model (example: Meta’s Llama 3.3 70B Instruct), you might run:
    To serve a large model (example: Meta’s Llama 3.3 70B Instruct), you might run:
    ```bash
    ```bash
-   vllm serve meta-llama/Llama-3.3-70B-Instruct --tensor-parallel-size=2
+   vllm serve meta-llama/Llama-3.3-70B-Instruct --tensor-parallel-size=4
    ```
    ```
    Adjust the model name and `--tensor-parallel-size` according to your hardware and parallelization needs.
    Adjust the model name and `--tensor-parallel-size` according to your hardware and parallelization needs.