Elo HeLLM: new project for ranking language models #12969
JohannesGaessler
started this conversation in
Show and tell
Replies: 1 comment 1 reply
-
If you think it could be useful to get more eyes on the project, feel free to add a link to it or to this discussion in the hot topics of the readme. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I started a new project called Elo HeLLM for evaluating model quality using the llama.cpp HTTP server. I intend to co-develop this project with the llama.cpp training code in order to have quality control since
llama-perplexity
is not suitable for determining whether a finetune is actually any good. By comparison, since the methods I'm using rely on the generation of tokens instead of evaluating the model on a pre-existing text the performance bottleneck is much more severe. So I intend to also look into improving the performance of batched inference using the server, particularly for multiple GPUs.Beta Was this translation helpful? Give feedback.
All reactions