-
I am having issues with converting qlora trained model(4bit) into a GGUF model using this script (https://github.com/ggerganov/llama.cpp/blob/master/examples/make-ggml.py) mainly because qlora trained model does not have a config.json. it has a adapter_config.json. on running this command i manually loaded the config.json of the llama 13b model which i'm using as a base model into the folder then the below error came up. is there a way to do it without config.json file? if not then how to correctly generate config.json file of a qlora trained model? |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 1 reply
-
you are dealing with a lora, which is an adapter for a model. if you want to use the lora, first convert it using |
Beta Was this translation helpful? Give feedback.
-
Thank you for the answer. I have a follow-up question. After converting LoRA weights to ggml using |
Beta Was this translation helpful? Give feedback.
you are dealing with a lora, which is an adapter for a model. if you want to use the lora, first convert it using
convert-lora-to-ggml.py
. then you can load the model and the lora. (it requires the base model).you can also merge the lora into the base model using the export-lora program.