Questions on chap05 10-llm-training-speed Multi gpu script #569
Answered
by
rasbt
STEVENTAN100
asked this question in
Q&A
-
Hi, @rasbt I've run |
Beta Was this translation helpful? Give feedback.
Answered by
rasbt
Mar 24, 2025
Replies: 1 comment 10 replies
-
Hi there, thanks for the command. And hm, that's weird! I know you probably don't want to train on the first GPU, but could you try
|
Beta Was this translation helpful? Give feedback.
10 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Oh I see. Thanks for clarifying. Yes, that's correct. I only focused on the
torchrun
code here to keep the code differences more minimal. Since most people usetorchrun
, and it is also the "PyTorch officially recommended way" I was planning to only recommend that as well moving forward. (Also, I didn't want to mix & match mp.spawn code for people who use torchrun, and I think it's just easier to let torchrun handle it).The README (https://github.com/rasbt/LLM…