Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

v0.8.4 #1136

Merged
merged 72 commits into from
Feb 24, 2024
Merged

v0.8.4 #1136

merged 72 commits into from
Feb 24, 2024

Conversation

kohya-ss
Copy link
Owner

No description provided.

kohya-ss and others added 30 commits November 23, 2023 22:17
remove missing import, and remove unused imports
fix broken import in svd_merge_lora script
Device support improvements (MPS)
Fix IPEX support and add XPU device to device_utils
…1061)

* Update train_util.py

Modifying to attempt enable multi GPU inference

* Update train_util.py

additional VRAM checking, refactor check_vram_usage to return string for use with accelerator.print

* Update train_network.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

remove sample image debug outputs

* Update train_util.py

* Update train_util.py

* Update train_network.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_network.py

* Update train_util.py

* Update train_network.py

* Update train_network.py

* Update train_network.py

* Cleanup of debugging outputs

* adopt more elegant coding

Co-authored-by: Aarni Koskela <[email protected]>

* Update train_util.py

Fix leftover debugging code
attempt to refactor inference into separate function

* refactor in function generate_per_device_prompt_list() generation of distributed prompt list

* Clean up missing variables

* fix syntax error

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* true random sample image generation

update code to reinitialize random seed to true random if seed was set

* true random sample image generation

* simplify per process prompt

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_util.py

* Update train_network.py

* Update train_network.py

* Update train_network.py

---------

Co-authored-by: Aarni Koskela <[email protected]>
@kohya-ss kohya-ss merged commit e69d341 into main Feb 24, 2024
2 checks passed
@FurkanGozukara
Copy link

Very nice work with highvram

How much speed up it leads and how much more VRAM it uses?

@kohya-ss
Copy link
Owner Author

It depends on --vae_batch_size. If you have have already using a maximum batch size for VRAM, I think the speed up is about 10-20%.

@FurkanGozukara
Copy link

vae_batch_size

I never used vae_batch_size. Only used regular batch size. Any difference?

@kohya-ss
Copy link
Owner Author

vae_batch_size is a batch size used in caching latents. The default value is 1, so it will be quite faster if increasing vae_batch_size as much as possible.

@FurkanGozukara
Copy link

vae_batch_size is a batch size used in caching latents. The default value is 1, so it will be quite faster if increasing vae_batch_size as much as possible.

haha i didnt know until all this time thank you so much for this amazing info

@kohya-ss
Copy link
Owner Author

Sorry the documents seem to lack this option... This option is a bit old and exists from v0.5.3. https://github.com/kohya-ss/sd-scripts/releases/tag/v0.5.3

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

8 participants