Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multi-gpu training degrades performance #95

Open
XiaoXuan42 opened this issue Dec 7, 2023 · 2 comments
Open

Multi-gpu training degrades performance #95

XiaoXuan42 opened this issue Dec 7, 2023 · 2 comments

Comments

@XiaoXuan42
Copy link

Hi, is there any caution with this lib when use multiple gpus in training? Like taking special attention to InnerBatchnorm, etc? I train two versions of the same network structure, one on one gpu while the other on multiple gpus and their performances differ a lot.

@maxxxzdn
Copy link

maxxxzdn commented Mar 5, 2024

Just curious, how severe is the degradation? Did you also adjust for batch size and learning rate when training on multiple GPUs?

@lrenaux-bdai
Copy link

I just posted #103 which could be what you're referring to

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants