-
Notifications
You must be signed in to change notification settings - Fork 7.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DeepSeek-R1 GGUF can't be loaded #3404
Comments
I am having the same issue. Version 3.6.1, Windows 10 Pro. |
I am having the same issue. Version 3.7.0, Windows 11 Pro. I tried many models of several people to run DeepSeek-R1, but without success so far. |
I suspect the chat template needs modification to accommodate the new output format. |
Same here. W11H, v3.7.0 |
Good morning all, Just to confirm, I tried some of the lower parameter versions of bartowski's Deepseek models (7B and 14B) but was receiving an error when trying to load the model into gpt4all. However, I tried loading in the larger model Hope this helps. |
Same here on Bazzite Linux |
I wonder if there is a compatible chat template generated for the 70B model which could be used for 14B or 32B even. |
I wonder if the Llama 8B distill works as well and it's only the QWEN models experiencing the issue. |
I have the same problem. The software is unable to directly open the downloaded gguf file. |
Same here, v3.7.0, W11H. Can't load 14B and 7B model:
|
Also possible solution here: |
the solution did not work for me |
Same problem here. v3.7.0, Win11. Can't load 7B, 8B, 14B DeepSeek-R1 Llama and Qwen models. |
Same problem, Win 11 |
PLEASE MAKE DEEPSEEK compatible to GPT4All, As I had tried all GGFUF of Deepseek, but none are working. LMStudio is working super fast and completely fine with Deepseek, so make it compatible too. |
This is not an issue, GPT4ALL docs clearly say custom models are not supported OOTB and it is up to user to additional configuration to work. This issue should close. |
A pr was merged into the repo a few hours ago that fixes this, but if you want to get it working right now use this chat template which is based on that pr: {%- if not add_generation_prompt is defined %}
{%- set add_generation_prompt = false %}
{%- endif %}
{%- if messages[0]['role'] == 'system' %}
{{- messages[0]['content'] }}
{%- endif %}
{%- for message in messages %}
{%- if message['role'] == 'user' %}
{{- '<|User|>' + message['content'] }}
{%- endif %}
{%- if message['role'] == 'assistant' %}
{%- set content = message['content'].split('</think>', 1) | last %}
{{- '<|Assistant|>' + content + '<|end▁of▁sentence|>' }}
{%- endif %}
{%- endfor -%}
{%- if add_generation_prompt %}
{{- '<|Assistant|>' }}
{%- endif %} |
Thanks you so much for this, it works now. But can it be possible or any plan of UI changes for Thinking Part, as it looks very good in DeepSeek online or LM Studio, the THINKING-REASONING PART should look beautifully in small box with small font with brain ICON blinking. |
Are you using this model- https://huggingface.co/bartowski/DeepSeek-R1-Distill-Llama-8B-GGUF ? |
Yes, Brother its Llama with 8B_Q4_K_M variant, its working super flawlessly. But Just now, I came to know they release new update 3.8, so its supported natively in app now, without any custom modification, as earlier. If you still face any issue in running it, just ping me out, I will be glad to help you in every way possible. TS |
Thanks you so much for this, it works now. |
Thanks, Debanna, I will give it a go.
…On Fri, Jan 31, 2025, 6:27 AM Debanna Das ***@***.***> wrote:
{%- if not add_generation_prompt is defined %}
{%- set add_generation_prompt = false %}
{%- endif %}
{%- if messages[0]['role'] == 'system' %}
{{- messages[0]['content'] }}
{%- endif %}
{%- for message in messages %}
{%- if message['role'] == 'user' %}
{{- '<|User|>' + message['content'] }}
{%- endif %}
{%- if message['role'] == 'assistant' %}
{%- set content = message['content'].split('', 1) | last %}
{{- '<|Assistant|>' + content + '<|end▁of▁sentence|>' }}
{%- endif %}
{%- endfor -%}
{%- if add_generation_prompt %}
{{- '<|Assistant|>' }}
{%- endif %}
*Thanks you so much for this, it works now. But can it be possible or any
plan of UI changes for Thinking Part, as it looks very good in DeepSeek
online or LM Studio, the THINKING-REASONING PART should look beautifully in
small box with small font with brain ICON blinking.*
Are you using this model-
https://huggingface.co/bartowski/DeepSeek-R1-Distill-Llama-8B-GGUF ?
Yes, Brother its Llama with *8B_Q4_K_M* variant, its working super
flawlessly.
But Just now, I came to know they *release new update 3.8,* so its
supported natively in app now, without any custom modification, as earlier.
If you still face any issue in running it, just ping me out, I will be
glad to help you in every way possible.
TS
Thanks you so much for this, it works now.
—
Reply to this email directly, view it on GitHub
<#3404 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AXDG6JMBM5Q4U2KRVB7OJDT2NNMYHAVCNFSM6AAAAABVTZQXUOVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDMMRWHE4DAOBXG4>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
I encounter the error 'item at index 3 is not a prompt' with all models in the reasoning tab. If I ask just one question, it works, but if I ask another one, I get that error. I should mention that I was using it with LocalDocs; apparently, that’s when it refuses to respond until the model is reloaded. |
@QohoZ : me too, but I think it's better if you open a separate issue. The problem you're describing is not related to this one |
Bug Report
GPT4All downloads the DeepSeek-R1 library, but receive an error attempting to load the model for chat.
Steps to Reproduce
Expected Behavior
GGuf file should load as with all of the others. Ollama loads the same file fine.
Your Environment
I notice in the Chat Template settings an error is present:
Syntax error: 1:267: error: This feature has not been supported yet
The text was updated successfully, but these errors were encountered: