-
Notifications
You must be signed in to change notification settings - Fork 161
invalid model file (bad magic [got 0x67676d66 want 0x67676a74]) #58
Comments
@abdeladim-s thanks, I just came to post that one has to download tokenizer as you pointed out (#5). I actually did, but then I get What does it mean by old file! Actually downloaded the most recent model.bin file from that link ([gpt4all-lora-quantized-ggml.bin] 05-Apr-2023 13:07 4G). Now, I am wondering how should I fix this to get the model working. |
I couldn't fix it to work, so I redownload the converted model: https://huggingface.co/LLukas22/gpt4all-lora-quantized-ggjt. I am trying this on my server with 2 core and 8GB of ram (I know it is the limit), and i tried to bring down temperature and ease up some of the parameter, yet it is stalling! Typically how fast should I expect this to run on such server? #Load the model #Generate result=model.generate(prompt,n_predict=50,temp=0, top_k = 3, top_p = 0.950000 ,repeat_last_n = 64, repeat_penalty = 1.100000) is there any hyperparameter to fix it to work faster? |
@qaiwiz the spec you are using is very low, you should have a quad core CPU at least. |
Here is the system config: |
Here is the output: llama_print_timings: load time = 71340.45 ms
|
I couldn't fix it to work, so I redownload the converted model: https://huggingface.co/LLukas22/gpt4all-lora-quantized-ggjt. |
guys you borked the lama again?
Was working until i did a git pull today. So, whats going on? How do you convert to the right magic?, We (GPT4ALL-UI) just recently converted all models and uploaded to the hf but now they are dead... Issue: ParisNeo/lollms-webui#96 |
@andzejsp am facing the same issue as well :/ , just tried it now with latest model and it doesn't work |
in my case its working with ggml-vicuna-13b-4bit-rev1.bin model, not sure why the other model died... |
@andzejsp can you give me a download link to it if you have so i can try it ? |
|
@andzejsp We didn't touch anything, we didn't push any updates since a week now. You can take a look at the commits history. |
I am working on linux debian 11, and after pip install and downloading a most recent mode: gpt4all-lora-quantized-ggml.bin I have tried to test the example but I get the following error:
./gpt4all-lora-quantized-ggml.bin: invalid model file (bad magic [got 0x67676d66 want 0x67676a74])
you most likely need to regenerate your ggml files
the benefit is you'll get 10-100x faster load times
see ggml-org/llama.cpp#91
use convert-pth-to-ggml.py to regenerate from original pth
use migrate-ggml-2023-03-30-pr613.py if you deleted originals
llama_init_from_file: failed to load model
I tried this: pyllamacpp-convert-gpt4all ./gpt4all-lora-quantized-ggml.bin ./llama_tokenizer ./gpt4all-converted.bin but I am not sure where the tokenizer is stored!
The text was updated successfully, but these errors were encountered: