-
-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Usage]: Trying to add codeshell 7b model, but got an error #11451
Comments
Is this a custom model? Please format your code properly using code blocks as it's difficult to read. |
What is the problem that you're running into? Please provide more context, not just an error message. |
I try to add a new model codeshell7b to vllm. I built codeshell.py, but encountered this error after running it. I don't know how to modify it. |
Do you get this error before vLLM finishes starting up, or only during inference? |
only during inference,can load model
|
An error occurs only when executing the following command:
|
I see that you're using an older version of vLLM. Can you try upgrading vLLM to make sure your dependencies are up-to-date? |
No, my environment does not support newer vllm. Could you please try to see if there is any problem with my code?
|
Can you run |
|
For vLLM 0.5.4, you should have PyTorch 2.4.0. See https://github.com/vllm-project/vllm/blob/v0.5.4/requirements-cuda.txt |
But other integrated models can run normally, such as qwen. |
I rule out incompatibilities in the dependencies first, before looking further into the code. Otherwise, it would be difficult for me to recreate your environment - how could I help you then? |
Since my device is not NVIDIA, the torch in my environment supports this version of vllm. Please help me check the code. The transformer code of codeshell7b is referenced at https://github.com/WisdomShell/codeshell/blob/09d1adc88ccada1a92924c69ece0cf0e73899b1b/model/modeling_codeshell.py |
model path:https://huggingface.co/WisdomShell/CodeShell-7B-Chat#model-details;I look forward to your help in solving the above problems. Thank you very much. |
When trying to run this model in the latest version of vLLM, I found that there is no |
OK, after reusing some of the code from the existing models, I'm able to run the model on my branch using the latest version of vLLM (https://github.com/DarkLight1337/vllm/tree/codeshell) without crashing. However the output is still garbled, probably need to step through the results with a debugger and compare with HF to see where it goes wrong. |
Thank you very much for your help. I am also having the garbled problem now. I will try to modify it first. |
Do you have any new progress on garbled characters? |
No, I thought you were looking into this. I am working on other PRs. |
Refer to #11681 |
Your current environment
The output of `python collect_env.py`
Model Input Dumps
No response
🐛 Describe the bug
####codeshell.py
###but running error
###error:
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: