You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+43-39Lines changed: 43 additions & 39 deletions
Original file line number
Diff line number
Diff line change
@@ -249,6 +249,49 @@ ControlNet-LLLite, a novel method for ControlNet with SDXL, is added. See [docum
249
249
250
250
## Change History
251
251
252
+
### Feb 24, 2024 / 2024/2/24: v0.8.4
253
+
254
+
- The log output has been improved. PR [#905](https://github.com/kohya-ss/sd-scripts/pull/905) Thanks to shirayu!
255
+
- The log is formatted by default. The `rich` library is required. Please see [Upgrade](#upgrade) and update the library.
256
+
- If `rich` is not installed, the log output will be the same as before.
257
+
- The following options are available in each training script:
258
+
-`--console_log_simple` option can be used to switch to the previous log output.
259
+
-`--console_log_level` option can be used to specify the log level. The default is `INFO`.
260
+
-`--console_log_file` option can be used to output the log to a file. The default is `None` (output to the console).
261
+
- The sample image generation during multi-GPU training is now done with multiple GPUs. PR [#1061](https://github.com/kohya-ss/sd-scripts/pull/1061) Thanks to DKnight54!
262
+
- The support for mps devices is improved. PR [#1054](https://github.com/kohya-ss/sd-scripts/pull/1054) Thanks to akx! If mps device exists instead of CUDA, the mps device is used automatically.
263
+
- The `--new_conv_rank` option to specify the new rank of Conv2d is added to `networks/resize_lora.py`. PR [#1102](https://github.com/kohya-ss/sd-scripts/pull/1102) Thanks to mgz-dev!
264
+
- An option `--highvram` to disable the optimization for environments with little VRAM is added to the training scripts. If you specify it when there is enough VRAM, the operation will be faster.
265
+
- Currently, only the cache part of latents is optimized.
266
+
- The IPEX support is improved. PR [#1086](https://github.com/kohya-ss/sd-scripts/pull/1086) Thanks to Disty0!
267
+
- Fixed a bug that `svd_merge_lora.py` crashes in some cases. PR [#1087](https://github.com/kohya-ss/sd-scripts/pull/1087) Thanks to mgz-dev!
268
+
- DyLoRA is fixed to work with SDXL. PR [#1126](https://github.com/kohya-ss/sd-scripts/pull/1126) Thanks to tamlog06!
269
+
- The common image generation script `gen_img.py` for SD 1/2 and SDXL is added. The basic functions are the same as the scripts for SD 1/2 and SDXL, but some new features are added.
270
+
- External scripts to generate prompts can be supported. It can be called with `--from_module` option. (The documentation will be added later)
271
+
- The normalization method after prompt weighting can be specified with `--emb_normalize_mode` option. `original` is the original method, `abs` is the normalization with the average of the absolute values, `none` is no normalization.
272
+
- Gradual Latent Hires fix is added to each generation script. See [here](./docs/gen_img_README-ja.md#about-gradual-latent) for details.
- Fixed a bug that the training crashes when `--fp8_base` is specified with `--save_state`. PR [#1079](https://github.com/kohya-ss/sd-scripts/pull/1079) Thanks to feffy380!
@@ -319,45 +362,6 @@ network_multiplier = -1.0
319
362
```
320
363
321
364
322
-
### Jan 17, 2024 / 2024/1/17: v0.8.1
323
-
324
-
- Fixed a bug that the VRAM usage without Text Encoder training is larger than before in training scripts for LoRA etc (`train_network.py`, `sdxl_train_network.py`).
325
-
- Text Encoders were not moved to CPU.
326
-
- Fixed typos. Thanks to akx! [PR #1053](https://github.com/kohya-ss/sd-scripts/pull/1053)
- Diffusers, Accelerate, Transformers and other related libraries have been updated. Please update the libraries with [Upgrade](#upgrade).
335
-
- Some model files (Text Encoder without position_id) based on the latest Transformers can be loaded.
336
-
-`torch.compile` is supported (experimental). PR [#1024](https://github.com/kohya-ss/sd-scripts/pull/1024) Thanks to p1atdev!
337
-
- This feature works only on Linux or WSL.
338
-
- Please specify `--torch_compile` option in each training script.
339
-
- You can select the backend with `--dynamo_backend` option. The default is `"inductor"`. `inductor` or `eager` seems to work.
340
-
- Please use `--sdpa` option instead of `--xformers` option.
341
-
- PyTorch 2.1 or later is recommended.
342
-
- Please see [PR](https://github.com/kohya-ss/sd-scripts/pull/1024) for details.
343
-
- The session name for wandb can be specified with `--wandb_run_name` option. PR [#1032](https://github.com/kohya-ss/sd-scripts/pull/1032) Thanks to hopl1t!
344
-
- IPEX library is updated. PR [#1030](https://github.com/kohya-ss/sd-scripts/pull/1030) Thanks to Disty0!
345
-
- Fixed a bug that Diffusers format model cannot be saved.
Gradual Latent is a Hires fix that gradually increases the size of the latent. `gen_img.py`, `sdxl_gen_img.py`, and `gen_img_diffusers.py` have the following options.
461
+
462
+
-`--gradual_latent_timesteps`: Specifies the timestep to start increasing the size of the latent. The default is None, which means Gradual Latent is not used. Please try around 750 at first.
463
+
-`--gradual_latent_ratio`: Specifies the initial size of the latent. The default is 0.5, which means it starts with half the default latent size.
464
+
-`--gradual_latent_ratio_step`: Specifies the ratio to increase the size of the latent. The default is 0.125, which means the latent size is gradually increased to 0.625, 0.75, 0.875, 1.0.
465
+
-`--gradual_latent_ratio_every_n_steps`: Specifies the interval to increase the size of the latent. The default is 3, which means the latent size is increased every 3 steps.
466
+
467
+
Each option can also be specified with prompt options, `--glt`, `--glr`, `--gls`, `--gle`.
468
+
469
+
__Please specify `euler_a` for the sampler.__ Because the source code of the sampler is modified. It will not work with other samplers.
470
+
471
+
It is more effective with SD 1.5. It is quite subtle with SDXL.
0 commit comments