Commit Graph

183 Commits

Author SHA1 Message Date
bmaltais
a1f6438f7b Upgrade upgrade.ps1 script to fix reported issue:
https://github.com/bmaltais/kohya_ss/issues/165
2023-02-14 17:42:36 -05:00
bmaltais
261b6790ee Update tool 2023-02-12 07:02:05 -05:00
bmaltais
a49fb9cb8c 2023/02/11 (v20.7.2):
- ``lora_interrogator.py`` is added in ``networks`` folder. See ``python networks\lora_interrogator.py -h`` for usage.
        - For LoRAs where the activation word is unknown, this script compares the output of Text Encoder after applying LoRA to that of unapplied to find out which token is affected by LoRA. Hopefully you can figure out the activation word. LoRA trained with captions does not seem to be able to interrogate.
        - Batch size can be large (like 64 or 128).
    - ``train_textual_inversion.py`` now supports multiple init words.
    - Following feature is reverted to be the same as before. Sorry for confusion:
        > Now the number of data in each batch is limited to the number of actual images (not duplicated). Because a certain bucket may contain smaller number of actual images, so the batch may contain same (duplicated) images.
    - Add new tool to sort, group and average crop image in a dataset
2023-02-11 11:59:38 -05:00
bmaltais
e5f8ba559f Add server_port and inbrowser support
- to all gui scripts
2023-02-10 08:22:03 -05:00
bmaltais
56d171c55b Reverting changes to startup commands 2023-02-09 20:25:42 -05:00
bmaltais
7bc93821a0 2023/02/09 (v20.7.1)
- Caption dropout is supported in ``train_db.py``, ``fine_tune.py`` and ``train_network.py``. Thanks to forestsource!
        - ``--caption_dropout_rate`` option specifies the dropout rate for captions (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the image is trained with the empty caption. Default is 0 (no dropout).
        - ``--caption_dropout_every_n_epochs`` option specifies how many epochs to drop captions. If ``3`` is specified, in epoch 3, 6, 9 ..., images are trained with all captions empty. Default is None (no dropout).
        - ``--caption_tag_dropout_rate`` option specified the dropout rate for tags (comma separated tokens) (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the tag is removed from the caption. If ``--keep_tokens`` option is set, these tokens (tags) are not dropped. Default is 0 (no droupout).
        - The bulk image downsampling script is added. Documentation is [here](https://github.com/kohya-ss/sd-scripts/blob/main/train_network_README-ja.md#%E7%94%BB%E5%83%8F%E3%83%AA%E3%82%B5%E3%82%A4%E3%82%BA%E3%82%B9%E3%82%AF%E3%83%AA%E3%83%97%E3%83%88) (in Jpanaese). Thanks to bmaltais!
        - Typo check is added. Thanks to shirayu!
    - Add option to autolaunch the GUI in a browser and set the server_port. USe either `gui.ps1 --inbrowser --server_port 3456`or `gui.cmd -inbrowser -server_port 3456`
2023-02-09 19:17:24 -05:00
bmaltais
90c0d55457 2023/02/09 (v20.7.1)
- Caption dropout is supported in ``train_db.py``, ``fine_tune.py`` and ``train_network.py``. Thanks to forestsource!
        - ``--caption_dropout_rate`` option specifies the dropout rate for captions (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the image is trained with the empty caption. Default is 0 (no dropout).
        - ``--caption_dropout_every_n_epochs`` option specifies how many epochs to drop captions. If ``3`` is specified, in epoch 3, 6, 9 ..., images are trained with all captions empty. Default is None (no dropout).
        - ``--caption_tag_dropout_rate`` option specified the dropout rate for tags (comma separated tokens) (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the tag is removed from the caption. If ``--keep_tokens`` option is set, these tokens (tags) are not dropped. Default is 0 (no droupout).
        - The bulk image downsampling script is added. Documentation is [here](https://github.com/kohya-ss/sd-scripts/blob/main/train_network_README-ja.md#%E7%94%BB%E5%83%8F%E3%83%AA%E3%82%B5%E3%82%A4%E3%82%BA%E3%82%B9%E3%82%AF%E3%83%AA%E3%83%97%E3%83%88) (in Jpanaese). Thanks to bmaltais!
        - Typo check is added. Thanks to shirayu!
    - Add option to autolaunch the GUI in a browser and set the server_port. USe either `gui.ps1 --inbrowser --server_port 3456`or `gui.cmd -inbrowser -server_port 3456`
2023-02-09 19:17:17 -05:00
bmaltais
09d3a72cd8 Adding support for caption dropout 2023-02-07 20:58:35 -05:00
bmaltais
8d559ded18 * 2023/02/06 (v20.7.0)
- ``--bucket_reso_steps`` and ``--bucket_no_upscale`` options are added to training scripts (fine tuning, DreamBooth, LoRA and Textual Inversion) and ``prepare_buckets_latents.py``.
    - ``--bucket_reso_steps`` takes the steps for buckets in aspect ratio bucketing. Default is 64, same as before.
        - Any value greater than or equal to 1 can be specified; 64 is highly recommended and a value divisible by 8 is recommended.
        - If less than 64 is specified, padding will occur within U-Net. The result is unknown.
        - If you specify a value that is not divisible by 8, it will be truncated to divisible by 8 inside VAE, because the size of the latent is 1/8 of the image size.
    - If ``--bucket_no_upscale`` option is specified, images smaller than the bucket size will be processed without upscaling.
        - Internally, a bucket smaller than the image size is created (for example, if the image is 300x300 and ``bucket_reso_steps=64``, the bucket is 256x256). The image will be trimmed.
        - Implementation of [#130](https://github.com/kohya-ss/sd-scripts/issues/130).
        - Images with an area larger than the maximum size specified by ``--resolution`` are downsampled to the max bucket size.
    - Now the number of data in each batch is limited to the number of actual images (not duplicated). Because a certain bucket may contain smaller number of actual images, so the batch may contain same (duplicated) images.
    - ``--random_crop`` now also works with buckets enabled.
        - Instead of always cropping the center of the image, the image is shifted left, right, up, and down to be used as the training data. This is expected to train to the edges of the image.
        - Implementation of discussion [#34](https://github.com/kohya-ss/sd-scripts/discussions/34).
2023-02-06 11:04:07 -05:00
bmaltais
cbfc311687 Integrate new bucket parameters in GUI 2023-02-05 20:07:00 -05:00
bmaltais
2486af9903 Update to latest dev code of kohya_s. WIP 2023-02-05 14:16:53 -05:00
bmaltais
2626214f8a Add support for LoRA resizing 2023-02-04 11:55:06 -05:00
bmaltais
045750b46a v20.6.0
- Increase max LoRA rank (dim) size to 1024.
    - Update finetune preprocessing scripts.
        - ``.bmp`` and ``.jpeg`` are supported. Thanks to breakcore2 and p1atdev!
        - The default weights of ``tag_images_by_wd14_tagger.py`` is now ``SmilingWolf/wd-v1-4-convnext-tagger-v2``. You can specify another model id from ``SmilingWolf`` by ``--repo_id`` option. Thanks to SmilingWolf for the great work.
        - To change the weight, remove ``wd14_tagger_model`` folder, and run the script again.
        - ``--max_data_loader_n_workers`` option is added to each script. This option uses the DataLoader for data loading to speed up loading, 20%~30% faster.
        - Please specify 2 or 4, depends on the number of CPU cores.
        - ``--recursive`` option is added to ``merge_dd_tags_to_metadata.py`` and ``merge_captions_to_metadata.py``, only works with ``--full_path``.
        - ``make_captions_by_git.py`` is added. It uses [GIT microsoft/git-large-textcaps](https://huggingface.co/microsoft/git-large-textcaps) for captioning.
        - ``requirements.txt`` is updated. If you use this script, [please update the libraries](https://github.com/kohya-ss/sd-scripts#upgrade).
        - Usage is almost the same as ``make_captions.py``, but batch size should be smaller.
        - ``--remove_words`` option removes as much text as possible (such as ``the word "XXXX" on it``).
        - ``--skip_existing`` option is added to ``prepare_buckets_latents.py``. Images with existing npz files are ignored by this option.
        - ``clean_captions_and_tags.py`` is updated to remove duplicated or conflicting tags, e.g. ``shirt`` is removed when ``white shirt`` exists. if ``black hair`` is with ``red hair``, both are removed.
    - Tag frequency is added to the metadata in ``train_network.py``. Thanks to space-nuko!
        - __All tags and number of occurrences of the tag are recorded.__ If you do not want it, disable metadata storing with ``--no_metadata`` option.
2023-02-04 08:36:35 -05:00
bmaltais
20e62af1a6 Update to latest kohya_ss sd-script code 2023-02-03 14:40:03 -05:00
bmaltais
c8f4c9d6e8 Add support for lr_scheduler_num_cycles, lr_scheduler_power 2023-01-30 08:26:15 -05:00
bmaltais
2ec7432440 Fix issue 81:
https://github.com/bmaltais/kohya_ss/issues/81
2023-01-29 11:17:30 -05:00
bmaltais
d45a7abb46 Add reference to Linux docker port 2023-01-29 11:12:05 -05:00
bmaltais
bc8a4757f8 Sync with kohya 2023/01/29 update 2023-01-29 11:10:06 -05:00
bmaltais
a4957cfea7 Adding LoRA tutorial 2023-01-27 19:46:13 -05:00
bmaltais
202923b3ce Add support for --keep_token option 2023-01-27 07:33:44 -05:00
bmaltais
bf371b49bf Fix issue 71 2023-01-27 07:04:35 -05:00
bmaltais
c6f87f1c01 Create save dir if not exist 2023-01-26 21:05:36 -05:00
bmaltais
03bd2e9b01 Add TI training support 2023-01-26 16:22:58 -05:00
bmaltais
49bada0d25 Update default save model file format to safetensors 2023-01-22 18:21:09 -05:00
bmaltais
511361c80b - Add new tool to verify LoRA weights produced by the trainer. Can be found under "Dreambooth LoRA/Tools/Verify LoRA 2023-01-22 11:40:14 -05:00
bmaltais
2ca17f69dd v20.4.0:
Add support for `network_alpha` under the Training tab and support for `--training_comment` under the Folders tab.
2023-01-22 10:18:00 -05:00
bmaltais
fcad6bfd98 Add default for caption extension tools 2023-01-20 18:41:37 -05:00
bmaltais
f265197ba9 Remove unneeded arguments 2023-01-19 20:51:40 -05:00
bmaltais
6a85f15542 Remove arg that were replaced 2023-01-19 20:49:40 -05:00
bmaltais
3e4b0ca720 Fix requirements file issue 2023-01-19 15:56:15 -05:00
bmaltais
29ec8658a5
Merge pull request #59 from bmaltais/dev
v20.3.0 release
2023-01-19 15:48:18 -05:00
bmaltais
31a1c8a71a Merge kohya Jan 19 updates 2023-01-19 15:47:43 -05:00
bmaltais
cb953d716f Update 2023-01-17 17:54:38 -05:00
bmaltais
9f95c8d830 New gui bat file
Update README
2023-01-17 17:54:20 -05:00
bmaltais
a615bcb66c
Merge pull request #44 from bmaltais/dev
Update gui start instructions
2023-01-16 13:40:16 -05:00
bmaltais
7886dfe9c7 Update gui start instructions 2023-01-16 13:39:10 -05:00
bmaltais
56c75a13aa
Merge pull request #43 from bmaltais/dev
Merging latest code update from kohya
2023-01-16 13:35:09 -05:00
bmaltais
95b9ab7c4d Update README 2023-01-16 13:33:17 -05:00
bmaltais
bfb0d18d4c Update install instructions 2023-01-16 10:28:20 -05:00
bmaltais
567221549f Complete training code refactoring 2023-01-15 19:59:40 -05:00
bmaltais
123cf4e3c5 Move cxommon adv train parm to common gui 2023-01-15 15:03:04 -05:00
bmaltais
abccecb093 Move save_state and resume to common gui
Format code
2023-01-15 12:01:59 -05:00
bmaltais
6aed2bb402 Add support for new arguments:
- max_train_epochs
- max_data_loader_n_workers
Move some of the codeto  common gui library.
2023-01-15 11:05:22 -05:00
bmaltais
43116feda8 Add support for max token 2023-01-10 09:38:32 -05:00
bmaltais
42a3646d4a Update readme 2023-01-09 17:59:11 -05:00
bmaltais
11fbc63440 Fix issue with lora model loading 2023-01-09 17:22:42 -05:00
bmaltais
f1b595a141
Merge pull request #35 from bmaltais/dev
Add support for --output_name to GUI
2023-01-09 11:49:53 -05:00
bmaltais
dc5afbb057 Move functions to common_gui
Add model name support
2023-01-09 11:48:57 -05:00
bmaltais
fdb4508a62 - Add model output name parameter to dreambooth
- Simplify config load and save
- blue code
2023-01-09 10:14:40 -05:00
bmaltais
402cb51ec0 refactor Dreambooth gui code 2023-01-09 08:08:47 -05:00