Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: xyz grid support multi-dimensional search&replace #3502

Open
2 tasks done
SAC020 opened this issue Oct 22, 2024 · 2 comments
Open
2 tasks done

[Feature]: xyz grid support multi-dimensional search&replace #3502

SAC020 opened this issue Oct 22, 2024 · 2 comments
Labels
enhancement New feature or request

Comments

@SAC020
Copy link

SAC020 commented Oct 22, 2024

Issue Description

xyz search and replace with only x activated, works fine, i.e. it applies lora <000002000>, then <000004000>

Prompt is:

photo of man on the street

comic book style <lora:Comic_book_V2:1.5> ,

<lora:sacbf-dalle-flux-v3-colab-captions_000002000:1.5>
17:29:23-790923 INFO     XYZ grid start: images=8 grid=1 shape=8x1 cells=1 steps=320
17:29:23-792918 DEBUG    XYZ grid apply prompt:
                         "<lora:sacbf-dalle-flux-v3-colab-captions_000002000:1.5>"="<lora:sacbf-dalle-flux-v3-colab-capt
                         ions_000002000:1.5>"
Load network: C:\ai\automatic\models\Lora\Comic_book_V2.safetensors ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/211.0 MB -:--:--
17:29:24-741893 DEBUG    LoRA name="Comic_book_V2" type={'ModuleTypeLora'} keys=338
Load network: C:\ai\automatic\models\Lora\sacbf-dalle-flux-v3-colab-captions_000002000.safetensors ━━━━━━ 0.0/1… -:--:--
                                                                                                          MB
17:29:25-582644 DEBUG    LoRA name="sacbf-dalle-flux-v3-colab-captions_000002000" type={'ModuleTypeLora'} keys=494
17:29:25-604586 INFO     Load network: type=LoRA apply=['Comic_book_V2', 'sacbf-dalle-flux-v3-colab-captions_000002000']
                         te=[1.5, 1.5] unet=[[1.5, 1.5, 1.5], [1.5, 1.5, 1.5]] dims=[None, None] load=1.81
17:29:25-609572 INFO     Base: class=FluxPipeline
17:29:25-611567 DEBUG    Sampler: sampler=default class=FlowMatchEulerDiscreteScheduler: {'num_train_timesteps': 1000,
                         'shift': 3.0, 'use_dynamic_shifting': True, 'base_shift': 0.5, 'max_shift': 1.15,
                         'base_image_seq_len': 256, 'max_image_seq_len': 4096}
17:29:25-613561 DEBUG    Torch generator: device=cuda seeds=[2218383567]
17:29:25-615556 DEBUG    Diffuser pipeline: FluxPipeline task=DiffusersTaskType.TEXT_2_IMAGE batch=1/1x1 set={'prompt':
                         1, 'guidance_scale': 4, 'num_inference_steps': 40, 'output_type': 'latent', 'width': 1024,
                         'height': 1024, 'parser': 'Fixed attention'}
Progress  1.28it/s ████████████████████████████▉      88% 35/40 00:30 00:03 Base17:29:59-974877 DEBUG    Server: alive=True jobs=1 requests=2486 uptime=4176 memory=32.11/63.92
                         backend=Backend.DIFFUSERS state=idle
Progress  1.15it/s █████████████████████████████████ 100% 40/40 00:34 00:00 Base
17:30:05-284987 DEBUG    VAE decode: vae name="default" dtype=torch.bfloat16 device=cuda:0 upcast=False slicing=True
                         tiling=True latents shape=torch.Size([1, 16, 128, 128]) dtype=torch.bfloat16 device=cuda:0
                         time=1.939
17:30:05-528334 INFO     Processed: images=1 its=0.96 time=41.73 timers={'init': 1.81, 'callback': 6.77, 'pipeline':
                         37.67, 'decode': 2.21, 'post': 0.03} memory={'ram': {'used': 37.84, 'total': 63.92}, 'gpu':
                         {'used': 9.0, 'total': 23.99}, 'retries': 0, 'oom': 0}
17:30:05-530330 DEBUG    XYZ grid apply prompt:
                         "<lora:sacbf-dalle-flux-v3-colab-captions_000002000:1.5>"="<lora:sacbf-dalle-flux-v3-colab-capt
                         ions_000004000:1.5>"
Load network: C:\ai\automatic\models\Lora\Comic_book_V2.safetensors ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/211.0 MB -:--:--
17:30:05-926271 DEBUG    LoRA name="Comic_book_V2" type={'ModuleTypeLora'} keys=338
Load network: C:\ai\automatic\models\Lora\sacbf-dalle-flux-v3-colab-captions_000004000.safetensors ━━━━━━ 0.0/1… -:--:--
                                                                                                          MB
17:30:07-296869 DEBUG    LoRA name="sacbf-dalle-flux-v3-colab-captions_000004000" type={'ModuleTypeLora'} keys=494
17:30:07-316816 INFO     Load network: type=LoRA apply=['Comic_book_V2', 'sacbf-dalle-flux-v3-colab-captions_000004000']
                         te=[1.5, 1.5] unet=[[1.5, 1.5, 1.5], [1.5, 1.5, 1.5]] dims=[None, None] load=1.79
17:30:07-319808 INFO     Base: class=FluxPipeline
17:30:07-321803 DEBUG    Sampler: sampler=default class=FlowMatchEulerDiscreteScheduler: {'num_train_timesteps': 1000,
                         'shift': 3.0, 'use_dynamic_shifting': True, 'base_shift': 0.5, 'max_shift': 1.15,
                         'base_image_seq_len': 256, 'max_image_seq_len': 4096}
17:30:07-323797 DEBUG    Torch generator: device=cuda seeds=[2218383567]
17:30:07-324795 DEBUG    Diffuser pipeline: FluxPipeline task=DiffusersTaskType.TEXT_2_IMAGE batch=1/1x1 set={'prompt':
                         1, 'guidance_scale': 4, 'num_inference_steps': 40, 'output_type': 'latent', 'width': 1024,
                         'height': 1024, 'parser': 'Fixed attention'}
Progress  1.00s/it █████████████████████████████████ 100% 40/40 00:40 00:00 Base
17:30:51-763889 DEBUG    VAE decode: vae name="default" dtype=torch.bfloat16 device=cuda:0 upcast=False slicing=True
                         tiling=True latents shape=torch.Size([1, 16, 128, 128]) dtype=torch.bfloat16 device=cuda:0
                         time=1.897
17:30:51-973328 INFO     Processed: images=1 its=0.86 time=46.44 timers={'init': 1.79, 'callback': 6.73, 'pipeline':
                         42.47, 'decode': 2.15, 'post': 0.02} memory={'ram': {'used': 37.91, 'total': 63.92}, 'gpu':
                         {'used': 8.99, 'total': 23.99}, 'retries': 0, 'oom': 0}

Same prompt, same x, but now with y activated as well, x gets ignored; it "says" it will replace the string, but applies the same lora <000002000> over and over again.

17:43:39-329259 INFO     XYZ grid start: images=24 grid=1 shape=8x3 cells=1 steps=960
17:43:39-331254 DEBUG    XYZ grid apply prompt:
                         "<lora:sacbf-dalle-flux-v3-colab-captions_000002000:1.5>"="<lora:sacbf-dalle-flux-v3-colab-capt
                         ions_000002000:1.5>"
17:43:39-332251 DEBUG    XYZ grid apply prompt: "comic book style <lora:Comic_book_V2:1.5>"="comic book style
                         <lora:Comic_book_V2:1.5>"
Load network: C:\ai\automatic\models\Lora\Comic_book_V2.safetensors ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/211.0 MB -:--:--
17:43:39-848385 DEBUG    LoRA name="Comic_book_V2" type={'ModuleTypeLora'} keys=338
Load network: C:\ai\automatic\models\Lora\sacbf-dalle-flux-v3-colab-captions_000002000.safetensors ━━━━━━ 0.0/1… -:--:--
                                                                                                          MB
17:43:40-737007 DEBUG    LoRA name="sacbf-dalle-flux-v3-colab-captions_000002000" type={'ModuleTypeLora'} keys=494
17:43:40-757952 INFO     Load network: type=LoRA apply=['Comic_book_V2', 'sacbf-dalle-flux-v3-colab-captions_000002000']
                         te=[1.5, 1.5] unet=[[1.5, 1.5, 1.5], [1.5, 1.5, 1.5]] dims=[None, None] load=1.42
17:43:40-763936 INFO     Base: class=FluxPipeline
17:43:40-766928 DEBUG    Sampler: sampler=default class=FlowMatchEulerDiscreteScheduler: {'num_train_timesteps': 1000,
                         'shift': 3.0, 'use_dynamic_shifting': True, 'base_shift': 0.5, 'max_shift': 1.15,
                         'base_image_seq_len': 256, 'max_image_seq_len': 4096}
17:43:40-769920 DEBUG    Torch generator: device=cuda seeds=[3557369830]
17:43:40-770917 DEBUG    Diffuser pipeline: FluxPipeline task=DiffusersTaskType.TEXT_2_IMAGE batch=1/1x1 set={'prompt':
                         1, 'guidance_scale': 4, 'num_inference_steps': 40, 'output_type': 'latent', 'width': 1024,
                         'height': 1024, 'parser': 'Fixed attention'}
Progress  1.04s/it ██████▊                             20% 8/40 00:15 00:33 Base17:43:59-845513 DEBUG    Server: alive=True jobs=1 requests=2916 uptime=5016 memory=46.89/63.92
                         backend=Backend.DIFFUSERS state=idle
Progress  1.04s/it █████████████████████████████████ 100% 40/40 00:41 00:00 Base
17:44:30-375357 DEBUG    VAE decode: vae name="default" dtype=torch.bfloat16 device=cuda:0 upcast=False slicing=True
                         tiling=True latents shape=torch.Size([1, 16, 128, 128]) dtype=torch.bfloat16 device=cuda:0
                         time=5.088
17:44:30-501021 INFO     Processed: images=1 its=0.78 time=51.17 timers={'init': 1.43, 'callback': 7.24, 'pipeline':
                         44.13, 'decode': 5.56, 'post': 0.03} memory={'ram': {'used': 41.08, 'total': 63.92}, 'gpu':
                         {'used': 5.13, 'total': 23.99}, 'retries': 0, 'oom': 0}
**17:44:30-505010 DEBUG    XYZ grid apply prompt:
                         "<lora:sacbf-dalle-flux-v3-colab-captions_000002000:1.5>"="<lora:sacbf-dalle-flux-v3-colab-capt
                         ions_000004000:1.5>"**
17:44:30-506009 DEBUG    XYZ grid apply prompt: "comic book style <lora:Comic_book_V2:1.5>"="comic book style
                         <lora:Comic_book_V2:1.5>"
Load network: C:\ai\automatic\models\Lora\Comic_book_V2.safetensors ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/211.0 MB -:--:--
17:44:30-996210 DEBUG    LoRA name="Comic_book_V2" type={'ModuleTypeLora'} keys=338
Load network: C:\ai\automatic\models\Lora\sacbf-dalle-flux-v3-colab-captions_**000002000**.safetensors ━━━━━━ 0.0/1… -:--:--
                                                                                                          MB
17:44:32-052895 DEBUG    LoRA name="sacbf-dalle-flux-v3-colab-captions_000002000" type={'ModuleTypeLora'} keys=494
17:44:32-072841 INFO     Load network: type=LoRA apply=['Comic_book_V2', 'sacbf-dalle-flux-v3-colab-captions_000002000']
                         te=[1.5, 1.5] unet=[[1.5, 1.5, 1.5], [1.5, 1.5, 1.5]] dims=[None, None] load=1.56
17:44:32-077828 INFO     Base: class=FluxPipeline
17:44:32-079822 DEBUG    Sampler: sampler=default class=FlowMatchEulerDiscreteScheduler: {'num_train_timesteps': 1000,
                         'shift': 3.0, 'use_dynamic_shifting': True, 'base_shift': 0.5, 'max_shift': 1.15,
                         'base_image_seq_len': 256, 'max_image_seq_len': 4096}
17:44:32-081846 DEBUG    Torch generator: device=cuda seeds=[3557369830]
17:44:32-082815 DEBUG    Diffuser pipeline: FluxPipeline task=DiffusersTaskType.TEXT_2_IMAGE batch=1/1x1 set={'prompt':
                         1, 'guidance_scale': 4, 'num_inference_steps': 40, 'output_type': 'latent', 'width': 1024,
                         'height': 1024, 'parser': 'Fixed attention'}
Progress  1.13it/s █████████████████████████████████ 100% 40/40 00:35 00:00 Base
17:45:12-477798 DEBUG    VAE decode: vae name="default" dtype=torch.bfloat16 device=cuda:0 upcast=False slicing=True
                         tiling=True latents shape=torch.Size([1, 16, 128, 128]) dtype=torch.bfloat16 device=cuda:0
                         time=1.888
17:45:12-706216 INFO     Processed: images=1 its=0.95 time=42.20 timers={'init': 1.57, 'callback': 7.12, 'pipeline':
                         38.37, 'decode': 2.22, 'post': 0.03} memory={'ram': {'used': 37.78, 'total': 63.92}, 'gpu':
                         {'used': 9.11, 'total': 23.99}, 'retries': 0, 'oom': 0}
17:45:12-713196 DEBUG    XYZ grid apply prompt:
                         "<lora:sacbf-dalle-flux-v3-colab-captions_000002000:1.5>"="<lora:sacbf-flux-dalle-v1_epoch_25:0
                         .7>"
17:45:12-715163 DEBUG    XYZ grid apply prompt: "comic book style <lora:Comic_book_V2:1.5>"="comic book style
                         <lora:Comic_book_V2:1.5>"
Load network: C:\ai\automatic\models\Lora\Comic_book_V2.safetensors ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 0.0/211.0 MB -:--:--
17:45:13-119224 DEBUG    LoRA name="Comic_book_V2" type={'ModuleTypeLora'} keys=338
Load network: C:\ai\automatic\models\Lora\sacbf-dalle-flux-v3-colab-captions_000002000.safetensors ━━━━━━ 0.0/1… -:--:--
                                                                                                          MB
17:45:13-874205 DEBUG    LoRA name="sacbf-dalle-flux-v3-colab-captions_000002000" type={'ModuleTypeLora'} keys=494
17:45:13-897144 INFO     Load network: type=LoRA apply=['Comic_book_V2', 'sacbf-dalle-flux-v3-colab-captions_000002000']
                         te=[1.5, 1.5] unet=[[1.5, 1.5, 1.5], [1.5, 1.5, 1.5]] dims=[None, None] load=1.18
17:45:13-901134 INFO     Base: class=FluxPipeline
17:45:13-902130 DEBUG    Sampler: sampler=default class=FlowMatchEulerDiscreteScheduler: {'num_train_timesteps': 1000,
                         'shift': 3.0, 'use_dynamic_shifting': True, 'base_shift': 0.5, 'max_shift': 1.15,
                         'base_image_seq_len': 256, 'max_image_seq_len': 4096}
17:45:13-905123 DEBUG    Torch generator: device=cuda seeds=[3557369830]
17:45:13-907117 DEBUG    Diffuser pipeline: FluxPipeline task=DiffusersTaskType.TEXT_2_IMAGE batch=1/1x1 set={'prompt':
                         1, 'guidance_scale': 4, 'num_inference_steps': 40, 'output_type': 'latent', 'width': 1024,
                         'height': 1024, 'parser': 'Fixed attention'}

It happens both with "script xyz" and "global / extension xyz".

Version Platform Description

PS C:\ai\automatic> .\webui --debug --medvram
Using VENV: C:\ai\automatic\venv
17:53:40-393526 INFO Starting SD.Next
17:53:40-396520 INFO Logger: file="C:\ai\automatic\sdnext.log" level=DEBUG size=65 mode=create
17:53:40-397516 INFO Python: version=3.11.9 platform=Windows bin="C:\ai\automatic\venv\Scripts\Python.exe"
venv="C:\ai\automatic\venv"
17:53:40-608585 INFO Version: app=sd.next updated=2024-10-21 hash=cc4f4819 branch=dev
url=https://github.com/vladmandic/automatic/tree/dev ui=dev
17:53:41-322526 INFO Repository latest available 5c684cb 2024-10-17T04:21:39Z
17:53:41-331950 INFO Platform: arch=AMD64 cpu=Intel64 Family 6 Model 165 Stepping 5, GenuineIntel system=Windows
release=Windows-10-10.0.22631-SP0 python=3.11.9

Relevant log output

No response

Backend

Diffusers

UI

Standard

Branch

Dev

Model

Other

Acknowledgements

  • I have read the above and searched for existing issues
  • I confirm that this is classified correctly and its not an extension issue
@vladmandic vladmandic changed the title [Issue]: xyz search & replace: x not getting applied when adding y [Feature]: xyz grid support multi-dimensional search&replace Oct 22, 2024
@vladmandic vladmandic added the enhancement New feature or request label Oct 22, 2024
@SAC020
Copy link
Author

SAC020 commented Oct 24, 2024

Not sure why this is marked as an enhancement, xyz used to support multi-dimensional S/R since forever (I have been using it since forever)

@SAC020
Copy link
Author

SAC020 commented Nov 5, 2024

Hi! Checking if there's anything planned on this, it's would really help my process of testing lora combos. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants