From 6662aebc27a9afa5faf926bd57722fe0bc25bbbe Mon Sep 17 00:00:00 2001 From: wangerxiao <863579016@qq.com> Date: Wed, 22 Jan 2025 20:39:27 +0800 Subject: [PATCH] [Misc] Fix the error in the tip for the --lora-modules parameter Signed-off-by: wangerxiao <863579016@qq.com> --- vllm/entrypoints/openai/cli_args.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/entrypoints/openai/cli_args.py b/vllm/entrypoints/openai/cli_args.py index 35445449463e9..4df75a665bab9 100644 --- a/vllm/entrypoints/openai/cli_args.py +++ b/vllm/entrypoints/openai/cli_args.py @@ -117,7 +117,7 @@ def make_arg_parser(parser: FlexibleArgumentParser) -> FlexibleArgumentParser: "or JSON format. " "Example (old format): ``'name=path'`` " "Example (new format): " - "``{\"name\": \"name\", \"local_path\": \"path\", " + "``{\"name\": \"name\", \"path\": \"lora_path\", " "\"base_model_name\": \"id\"}``") parser.add_argument( "--prompt-adapters",