From fde8340550d5c5587a97eef69839d3e8a57a43c8 Mon Sep 17 00:00:00 2001 From: Yineng Zhang Date: Mon, 5 Aug 2024 18:06:06 +0800 Subject: [PATCH] docs: update README (#935) --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index c29a72e3954..1bb6f13d0f8 100644 --- a/README.md +++ b/README.md @@ -151,9 +151,9 @@ python -m sglang.launch_server --model-path meta-llama/Meta-Llama-3-8B-Instruct ``` python -m sglang.launch_server --model-path meta-llama/Meta-Llama-3-8B-Instruct --port 30000 --mem-fraction-static 0.7 ``` -- If you see out-of-memory errors during prefill for long prompts, try to enable chunked prefill. +- If you see out-of-memory errors during prefill for long prompts on a model that supports long context, consider using chunked prefill. ``` -python -m sglang.launch_server --model-path meta-llama/Meta-Llama-3-8B-Instruct --port 30000 --chunked-prefill-size 8192 +python -m sglang.launch_server --model-path meta-llama/Meta-Llama-3.1-8B-Instruct --port 30000 --chunked-prefill-size 8192 ``` - See [hyperparameter_tuning.md](docs/en/hyperparameter_tuning.md) on tuning hyperparameters for better performance. - Add `--nnodes 2` to run tensor parallelism on multiple nodes. If you have two nodes with two GPUs on each node and want to run TP=4, let `sgl-dev-0` be the hostname of the first node and `50000` be an available port.