Loading GGML converted qlora models #4018
Answered
by
KerfuffleV2
ragesh2000
asked this question in
Q&A
-
I have lora weights of a finetuned model (adapter_model.bin) and i created a ggml version of the file using the python file convert-lora-to-ggml.py and now i have the ggml_model.bin file. |
Beta Was this translation helpful? Give feedback.
Answered by
KerfuffleV2
Nov 12, 2023
Replies: 1 comment 11 replies
-
|
Beta Was this translation helpful? Give feedback.
11 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
You need to use the base (full) model with
-m
and the converted LoRA with--lora
. So it should look like-m my_full_model.gguf --lora ggml-adapter-model.bin
Note:
-l
is not the short form of--lora
, it is for setting logit bias.