-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue when running on multiple GPUs #1438
Comments
Hi @zyzhang1130 |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. |
face the same issue too. |
@younesbelkada can you please show me a simple example script, where either |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. |
Hi all. Would you like to give the new PPOv2Trainer a try? They should work well with multiple GPUs. Feel free to re-open the issue if you run into this problem again. |
hello , i want to know that if the PPOv2Trainer support multi machines and multi gpus , i use torchrun to run the python scripts , thanks ~ |
I used PPOTrainer in a setting with 2 GPUs. I got the following issue:
But upon checking,
query_tensors
,negative_qna
,rewards
are all oncuda:1
. So I don't know why it says I have something oncuda:0
,The text was updated successfully, but these errors were encountered: