Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Model is not being offloaded from VRAM #35

Open
nishithshowri006 opened this issue Oct 8, 2024 · 0 comments
Open

Model is not being offloaded from VRAM #35

nishithshowri006 opened this issue Oct 8, 2024 · 0 comments

Comments

@nishithshowri006
Copy link

nishithshowri006 commented Oct 8, 2024

I am trying to run the model in Jupyter notebook.

image

  1. In the above iteration I haven't initialized the model.

image
2. Now I run the cell the model is loaded and it is showing 6GB of vram occupied right.

image
3. Now when I run the cell again the vram usage is doubled.
4. In the consequent runs the model is not occupying more than 12GB but what's interesting thing I have observed is when I am running that inside a loop for suppose I want to create an Index for each file I have, I don't have any other option than do this but this is causing the model to give me vram issues. How do I remove them from vram, I tried torch cuda cache free, tried to delete the variable none isn't working for me. Can you please help or is there something I am doing wrongly ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant