You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Unfortunately, it's expected. You could try to use older versions of llama.cpp, before #3228 was merged.
MPI should be fixed in #3334, more context: #3752.
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
run 2 chunks of the model on the same CPU-GPU
Current Behavior
ERRORS:
GGML_ASSERT: llama.cpp:8672: false && "not implemented"
GGML_ASSERT: llama.cpp:5443: false && "not implemented"
Environment and Context
Macbook M1 Max 32GB
Steps to Reproduce
Failure Logs
output.txt
The text was updated successfully, but these errors were encountered: