Fine tune data format for llama2 + prompting after exporting LORA #4757
Replies: 2 comments
-
Oh and I have tried this with non-quantised llama2-7b. It's the same results. |
Beta Was this translation helpful? Give feedback.
-
I'm not sure about my following comments but perhaps they might help. My concern would be that llama might not understand the json format of the samples. It would instead expect the samples to be in the format it was trained on. It might be worth a shot to restructure the samples to follow that format and see if that helps. For example, something like this:
Since the samples are in json format it should be possible to write a conversion function for your existing samples. Also if you want to use |
Beta Was this translation helpful? Give feedback.
-
Hi all,
I'm using fine-tune on data where each entry looks like this:
I'm using the
<s>
between each entry and I'm running the fine tune via:I run until the loss rate is
< 0.02
or so, 20 ish epochs. But when I then run the model afterwards I seem to only get nonsense.For example, I'll prompt:
and get a response:
Any ideas on what Im doing wrong with the tuning?
Beta Was this translation helpful? Give feedback.
All reactions