Running same prompt with different hardware(CPU & Customer hardware) lead to different prompt response even with temp:0.0 value #12167
akapoor3518
started this conversation in
General
Replies: 1 comment 8 replies
-
@akapoor3518 Does it happen when you set the same integer random seed value? (-s or --seed parameter). |
Beta Was this translation helpful? Give feedback.
8 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I am running below prompt (with --temp 0.0) on CPU and my customer hardware and i expected same response. But i am getting different response. I will looking through Llama.cpp code and see why this is difference. It will help if you guys can also give your suggestion on this.
Below what i had run:
./build/bin/llama-cli -p "my cat name" -m ./models/tinyllama-vo-5m-para.gguf --device none -c 12288 --temp 0.0 --n-predict 4 --repeat-penalty 1.5 -b 1024 --top-k 50 --top-p 0.9 --repeat-last-n 5 --no-warmup
Prompt response
###########
my cat name was a little girl
With Customer Hardware only following operation: GGML_OP_NONE, GGML_OP_ADD, GGML_OP_SUB, GGML_OP_DIV, GGML_OP_MUL were offloaded to customer hardware, rest going to CPU
./build/bin/llama-cli -p "my cat name" -m ./models/tinyllama-vo-5m-para.gguf --device customer-hardware -c 12288 --temp 0.0 --n-predict 4 --repeat-penalty 1.5 -b 1024 --top-k 50 --top-p 0.9 --repeat-last-n 5 --no-warmup
Prompt response
###########
my cat name the sun was shining
Beta Was this translation helpful? Give feedback.
All reactions