-
Notifications
You must be signed in to change notification settings - Fork 80
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[LLaMA] uncaught error while running! details: [LayerImpl] Unknown Layer Properties count 2 #2782
Comments
cibot: Thank you for posting issue #2782. The person in charge will reply soon. |
@dyna-bytes Hi, Thank you for reporting issues |
Hi, @DonghakPark |
i think error caused by using nntrainer's multi-head attention instead of custom_multi_head_attention in below section. Have you tried building with {
layers.push_back(createLayer(
"multi_head_attention",
{withKey("name", "layer" + std::to_string(layer_id) + "_attention_out"),
withKey("num_heads", std::to_string(NUM_HEADS)),
withKey("max_timestep", std::to_string(MAX_SEQ_LEN)),
withKey("disable_bias", "true"),
withKey("input_layers", {query_name, key_name, value_name})}));
} one more, for running Application, you need the following bin file std::string weight_path = "./llama_fp16.bin";
g_model->load(weight_path); |
I also recommend trying out meson build (since the currently released version has been tested with meson build). To run your own LLaMA model, you need to prepare three additional files. |
FYI, nntrainer provides some guidelines to debug remotely on adb shell with lldb-server. |
Hi, thank you all for your sincere replies. But I still get the same error message even though I build with I think the build didn't get done correctly. This is my build sequence.
I got these files builded,
and I got these so files too.
I can also find libcustom_multi_head_attention_layer.so generated.
Please help me figure out or debug what's wrong during build sequence. Thank you for your time. |
Hi, I tried to run LLaMa on Galaxy Z FLIP5 (Android14) and got some errors.
terminating with uncaught exception of type std::runtime_error: collate_byname<char>::collate_byname failed to construct for ko_KR.UTF-8
I got above error when I first tried to execute LLaMA on nntrainer.
I assumed that error should be related with NDK compiler and just removed the locale setting line from Applications/LLaMA/jni/main.c
So the error message was disappeared, but I don't think this is a fundamental solution.
The real problem was creating multi_head_attention layer of embedding0 block.
transformer = createTransformerDecoder(i, "embedding0");
The error messaged was generated while runnig 'createAttentionLayer()'.
I added some print line to debug as below.
I got the below error message while executing llama but couldn't figure out any reason for Unkown Properties error.
Enviroment:
I used android-ndk-r22b with Ubuntu 20.04 on docker container.
You can download my environment settings on the below url.
https://hub.docker.com/repository/docker/jetpark/nntrainer/general
Please help me solve this problem. Thank you.
The text was updated successfully, but these errors were encountered: