-
-
Notifications
You must be signed in to change notification settings - Fork 423
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Issue]: DeepCache with cpu offload generates distorted images after first batch #2888
Closed
2 tasks done
Comments
i've implemented suggested workaround for now. |
frostyplanet
pushed a commit
to frostyplanet/inference
that referenced
this issue
Sep 14, 2024
test result shows deepcache should be loaded before cpu_offloading if not, may cause issues like horseee/DeepCache#23 vladmandic/automatic#2888 Signed-off-by: wxiwnd <[email protected]>
frostyplanet
pushed a commit
to frostyplanet/inference
that referenced
this issue
Sep 14, 2024
test result shows deepcache should be loaded before cpu_offloading if not, may cause issues like horseee/DeepCache#23 vladmandic/automatic#2888 Signed-off-by: wxiwnd <[email protected]>
frostyplanet
pushed a commit
to frostyplanet/inference
that referenced
this issue
Sep 21, 2024
test result shows deepcache should be loaded before cpu_offloading if not, may cause issues like horseee/DeepCache#23 vladmandic/automatic#2888 Signed-off-by: wxiwnd <[email protected]>
frostyplanet
pushed a commit
to frostyplanet/inference
that referenced
this issue
Sep 27, 2024
test result shows deepcache should be loaded before cpu_offloading if not, may cause issues like horseee/DeepCache#23 vladmandic/automatic#2888 Signed-off-by: wxiwnd <[email protected]>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Issue Description
Found a problem while testing DeepCache on dev branch. If you enable model cpu offload first image will be generated as expected (3x speed up, expected quality), but all next batches will be distorted and have 9x speed up. It does not happen with sequential cpu offload and with no offload. The problem exist with all tested pipelines. I was able to figure out that this is problem in DeepCache and created issue on it's github (#horseee/DeepCache#23) with recreation of it on kaggle.
There is also a possible temporary fix for this problem. If you move enabling and disabling deepcache_worker from model compilation to pipeline execution (enable before and disable after) you can fix this issue.
Example of first 2 generated batches:
Version Platform Description
Python 3.10.13 on Linux
Version: app=sd.next updated=2024-02-18 hash=be81d486 url=https://github.com/vladmandic/automatic.git/tree/dev
Latest published version: 9c12b74 2024-02-18T22:40:13Z
Platform: arch=x86_64 cpu= system=Linux release=6.6.17-1-lts python=3.10.13
AMD ROCm toolkit detected
ROCm agents detected: ['gfx1032', 'gfx90c']
ROCm agent used by default: idx=0 gpu=gfx1032 arch=navi2x
ROCm version detected: 6.0
Load packages: {'torch': '2.3.0.dev20240218+rocm6.0', 'diffusers': '0.26.3', 'gradio': '3.43.2'}
Backend.DIFFUSERS compute=rocm device=cuda attention="Scaled-Dot-Product" mode=no_grad
Device: device=AMD Radeon RX 6600M n=1 hip=6.0.32830-d62f6a171
Relevant log output
Backend
Diffusers
Branch
Dev
Model
SD 1.5
Acknowledgements
The text was updated successfully, but these errors were encountered: