-
Notifications
You must be signed in to change notification settings - Fork 10.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cmake file always assumes AVX2 support #1583
Comments
check this #809 |
This issue is causing an issue downstream on "llama-cpp-python" where we cant build a python binding on non supported AVX2 machines that require cuBLAS support. |
As per my now-closed issue #1654 (currently closed by me because I figured out the workaround and wasn't sure if default configuration qualified as a "bug"), it assumes a bunch of other extensions as well: AVX, F16C, and FMA. It took me a while to figure out what the flags to disable them were and then add them one by one until it finally worked. |
Confirm the basically blocks installing llama-cpp-python on a machine without AVX2 available. |
Anyone have a straight forward way to get the combo of CUDA + no AVX2 to work? |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
When running
cmake
the default configuration sets AVX2 to be ON even when the current cpu does not support it.AVX vs AVX2 is handled correctly in the plain makefile.
For cmake, the AVX2 has to be turned off via
cmake -DLLAMA_AVX2=off .
for the compiled binary to work on AVX-only system.Can we make the cmake file smarter about whether to enable or disable AVX2 by looking at the current architecture?
The text was updated successfully, but these errors were encountered: