-
Notifications
You must be signed in to change notification settings - Fork 10.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Model not loading - nothing happens #1543
Comments
The file format has changed for Q4_0, Q4_1 and Q8_0. The files on that repo are too old. What you're seeing is the program silently crashing. True, that's not something desirable, maybe an issue on its own. Q5_0 and Q5_1 models should still work as before. It was changed here: #1508 |
Interesting, I was using the model "ggml-vic7b-uncensored-q4_0.bin", which has been updated for the new quantization. Apparently not correctly, because after what you said I tried the q5 version and it works like a charm. Happenend with the 7b and 13b model. Anyways, thank you for your help. |
Yeah, it was updated, but then the format changed again :) |
I just checked out an older version that dates back to september 11 so that i can use older GGML models :) |
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
Trying to get llama.cpp to run on windows 10
Current Behavior
Generally everything seems to work, but the model does not load. After it tries to load the model, it just exits after a few seconds and that is it. No error, message, nothing. RAM gets loaded for a few seconds, same with CPU and GPU.
I have tried every binary available, always with the same outcome. Also the two models I have tried are: ggml-vicuna-7b-1.1 and ggml-vicuna-13b-1.1
If I try for example ".\main.exe --help" the output is correct. So it seems to work in principle, also my GPU gets detected if the binary is with clblast.
![grafik](https://private-user-images.githubusercontent.com/102863031/239707395-aa3742b4-b6ee-486b-abfe-99087cfcef23.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk0NDIwODUsIm5iZiI6MTczOTQ0MTc4NSwicGF0aCI6Ii8xMDI4NjMwMzEvMjM5NzA3Mzk1LWFhMzc0MmI0LWI2ZWUtNDg2Yi1hYmZlLTk5MDg3Y2ZjZWYyMy5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjUwMjEzJTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI1MDIxM1QxMDE2MjVaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT00MGNiM2VjZmRhZTE1Y2ZkMDMxNmJkNTZmMzAwNTBmYjEzNzMxNzExMTJhYWEyMzZiZWQxOTc2NDFkMzYxMDA0JlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCJ9.hT0naxiUANdaXhPvqZt5ZRkLX8KHx5f02A7glLkOuBE)
Below I attached the console input and output. I am really out of words. Hours of googling did not seem to yield anything, because there is apparently nobody with that problem.
Environment and Context
The text was updated successfully, but these errors were encountered: