-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
A CUDA error is thrown when using llama_local
- ERR_PNPM_RECURSIVE_RUN_FIRST_FAIL, exit code 3221226505
#2080
Comments
Hello @ropstah! Welcome to the ai16z community. Thank you for opening your first issue; we appreciate your contribution. You are now a ai16z contributor! |
I am also experiencing this issue, and have been trying to solve it for a few days now without any luck. The error is thrown when calling
This might be another bug, but when using the cpu or vulkan the model is stuck in an endless loop. I saw this PR which fixes it - but the same fix doesn't work for the issue mentioned here. |
The local model should not have issues like this specially on CUDA. |
I also got the
And I only have one C: drive, so not sure how this path shows up 🤔 |
Same error here with:
Nothing really configured, I've just copied/pasted the .env.example to .env. Maybe we miss some configuration there ? I'll dig a bit. |
Anyone able to resolve this issue? |
Is this fixed? |
Describe the bug
A CUDA error is thrown when using
llama_local
To Reproduce
Windows 10
wsl2
node -v
=v23.5.0
python --version
=Python 3.12.8
followed Quick Start:
v0.1.7
pnpm install --no-frozen-lockfile
pnpm build
.env
setup forllama_local
:XAI_MODEL=meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
as described heremodelProvider
tollama_local
in trump.character.jsonpnpm start --character="characters/trump.character.json"
pnpm start:client
hi
in chatExpected behavior
An indication of what the error is exactly. I also changing model configuration but that doesn't seem to be picked up. I'm at a loss.
Screenshot

Additional context
The text was updated successfully, but these errors were encountered: