You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for sharing your code and detailed documentation. I am currently using your code and have encountered a few questions regarding the GPU usage.
What is the approximate GPU memory requirement for fine-tuning?
Is it possible to train the model using a single GPU? If so, what settings should I be aware of?
I am currently using multiple RTX 3090 GPUs but encountering out-of-memory (OOM) errors. Is there any way to resolve this issue? If so, how should I set it up?
Interestingly, I can successfully run test.py using a single RTX 3090 GPU, but it fails with multiple RTX 3090 GPUs. Do you have any suggestions on how to resolve this?
Thank you very much for your time and assistance!
The text was updated successfully, but these errors were encountered:
As for the failure of running the test with multiple GPUs, I think it may be because the communication between multiple GPUs takes up extra memory. You can reduce the batch size or num_beams appropriately.
Thank you for sharing your code and detailed documentation. I am currently using your code and have encountered a few questions regarding the GPU usage.
Thank you very much for your time and assistance!
The text was updated successfully, but these errors were encountered: