Replies: 2 comments
-
Not directly related to your synthetic minimal model, but maybe of interest: Regarding your error message: You'd have to single step for the error, I'd start at init_mapping() in llama.cpp. |
Beta Was this translation helpful? Give feedback.
-
It's probably throwing when trying to find the newline token: |
Beta Was this translation helpful? Give feedback.
-
As a side-project, I'm attempting to create a minimal GGUF model that can successfully be loaded by llama.cpp (through llama-cpp-python) - very much related to this question: #5038
The code that I'd like to successfully run is:
or, somewhat similar from llama.cpp directly with
./llama.cpp/main -m minimal.gguf -I
I do not expect actual inference to work, I'd just like for it to be able to load the model.
Looking at
gguf-py/gguf-py/examples/writer.py
, I modified it as follows:That code produces a GGUF file, that if dumped with
llama.cpp/gguf-py/scripts/gguf-dump.py
looks like this:(I straight-up took all KV values from TheBlokes tinyllm model)
My issue is that when I try to load that model up with llama.cpp, I get this:
I can't seem to locate the source of the
error loading model: unordered_map::at: key not found
message, perhaps because I haven't touched C++ in a really, really long time.Does anyone have any pointers as to how to progress on this goal? Is there a better way to build a barebones minimal model?
Full disclosure, details on the internals of actual models elude me - this here is me embarking on the journey to learn more about these internals ;)
Beta Was this translation helpful? Give feedback.
All reactions