Commit Graph

  • c61ad5f8f9 Fix for LoCon sliders visibility bmaltais 2023-03-03 10:21:24 -0500
  • 4812e441a0
    Merge pull request #283 from bmaltais/dev bmaltais 2023-03-03 07:45:04 -0500
  • 29bb8599bb Fix issue 278 bmaltais 2023-03-03 07:41:44 -0500
  • e482ff0b1e
    Merge pull request #281 from bmaltais/dev bmaltais 2023-03-03 07:11:51 -0500
  • d30abe5491 Fix issue 277 bmaltais 2023-03-03 07:11:15 -0500
  • 76acf2a200
    Merge pull request #273 from bmaltais/dev bmaltais 2023-03-02 20:39:33 -0500
  • 4c762e3abb Upgrade Gradio Fix Gradio issues Fix missing LoCon new GUI bmaltais 2023-03-02 20:39:07 -0500
  • 26fb489a72
    Merge pull request #271 from bmaltais/dev bmaltais 2023-03-02 17:52:02 -0500
  • 482d7834d1 Emergency fix bmaltais 2023-03-02 17:51:17 -0500
  • 78bffe9817 Remove locon folder bmaltais 2023-03-02 15:35:42 -0500
  • 0217a30563
    Merge pull request #269 from bmaltais/dev bmaltais 2023-03-02 15:33:44 -0500
  • 0253d46411 Adding locon to .gitignore bmaltais 2023-03-02 15:33:01 -0500
  • 64a084e74a
    Merge pull request #268 from bmaltais/dev bmaltais 2023-03-02 15:02:01 -0500
  • 962628c89a Fix ToC links bmaltais 2023-03-02 15:00:55 -0500
  • 6edc53ae3e
    Merge pull request #267 from bmaltais/dev bmaltais 2023-03-02 14:36:48 -0500
  • c926c9d877 Update Readme bmaltais 2023-03-02 14:36:07 -0500
  • 6105eb0279 Update readme bmaltais 2023-03-02 14:25:11 -0500
  • 11b3955032
    Merge branch 'LoCon' into dev bmaltais 2023-03-02 07:53:01 -0500
  • 384009d4eb
    Merge pull request #251 from Ki-wimon/master bmaltais 2023-03-02 07:09:16 -0500
  • 182080bb78 Run blue bmaltais 2023-03-01 19:24:11 -0500
  • 5498539fda Fix typos bmaltais 2023-03-01 19:20:05 -0500
  • 078bd8d768 Add typos checker bmaltais 2023-03-01 19:14:48 -0500
  • f6b261de52
    Merge pull request #257 from bmaltais/dev bmaltais 2023-03-01 19:03:03 -0500
  • 7f0e5683c6 v21.0.1 bmaltais 2023-03-01 19:02:04 -0500
  • 1e3055c895 Update tensorboard bmaltais 2023-03-01 13:14:47 -0500
  • 5bfd478d32
    update description to LoCon training Ki-wimon 2023-03-01 20:27:23 +0800
  • 88a49df47e
    set inital value to LoCon conv parameters Ki-wimon 2023-03-01 18:57:02 +0800
  • cdb8bb1182
    update code of cloning locon scripts Ki-wimon 2023-03-01 13:07:38 +0800
  • 6bcd52c9cc
    update new locon args support Ki-wimon 2023-03-01 12:22:11 +0800
  • c07e3bba76
    add new LoCon args Ki-wimon 2023-03-01 12:19:18 +0800
  • d76fe7d4e0
    LoCon script auto upgrade feature Ki-wimon 2023-02-28 22:58:35 +0800
  • b1fb87a9e1 Merging PR into LoCon branch bmaltais 2023-02-28 07:45:42 -0500
  • dfd155a8e1 Undo LoCon commit bmaltais 2023-02-28 07:37:19 -0500
  • 04f0f0cf4f
    Merge pull request #245 from Ki-wimon/master bmaltais 2023-02-28 07:04:27 -0500
  • f6bec77eaa
    Merge branch 'dev' into master bmaltais 2023-02-28 07:04:08 -0500
  • c32a99dad5
    Update lora_gui.py Ki-wimon 2023-02-28 01:38:05 +0800
  • 6e664f1176
    support locon Ki-wimon 2023-02-28 01:16:23 +0800
  • 9d2e3f85a2 Add tensorboard support bmaltais 2023-02-26 19:49:22 -0500
  • 74cce23116
    Merge pull request #238 from bmaltais/dev bmaltais 2023-02-26 15:14:02 -0500
  • 5e267b23af Adding activate script bmaltais 2023-02-26 15:12:50 -0500
  • 6b5d6303cc Update bmaltais 2023-02-26 15:11:21 -0500
  • f213b15014 Updates bmaltais 2023-02-24 20:37:51 -0500
  • 3b93266aae Bug: Fix issue: https://github.com/bmaltais/kohya_ss/issues/231 bmaltais 2023-02-24 07:30:37 -0500
  • 8775667fc7 Updates bmaltais 2023-02-23 21:48:45 -0500
  • c7e99eb54b
    Merge pull request #227 from bmaltais/dev bmaltais 2023-02-23 19:26:16 -0500
  • fe4558633d Fix date bmaltais 2023-02-23 19:25:51 -0500
  • df6092a52b Update readme bmaltais 2023-02-23 19:25:09 -0500
  • 60ad22733c Update to latest code version bmaltais 2023-02-23 19:21:30 -0500
  • 49bfd3a618
    Merge pull request #223 from bmaltais/dev bmaltais 2023-02-22 20:33:41 -0500
  • bf0344ba9e Adding GUI support for new features bmaltais 2023-02-22 20:32:57 -0500
  • 2a5fb346d5 Sinc to latest code update on sd-script bmaltais 2023-02-22 13:30:29 -0500
  • 34ab8448fb Fix issue where dadaptation code was pushed by mistake bmaltais 2023-02-20 08:26:45 -0500
  • 1807c548b5 Merge branch 'dev' of https://github.com/bmaltais/kohya_ss into dev bmaltais 2023-02-20 07:56:40 -0500
  • dfc9a8dd40 Fix issue with save config bmaltais 2023-02-20 07:56:24 -0500
  • 39ac6b0086
    Merge pull request #209 from bmaltais/dev bmaltais 2023-02-19 20:16:51 -0500
  • 611a0f3e76
    Merge branch 'master' into dev bmaltais 2023-02-19 20:16:44 -0500
  • 758bfe85dc Adding support for Lion optimizer in gui bmaltais 2023-02-19 20:13:03 -0500
  • bb57c1a36e Update code to latest sd-script version bmaltais 2023-02-19 06:50:33 -0500
  • 48122347a3
    Merge pull request #189 from bmaltais/LR-Free bmaltais 2023-02-17 19:18:39 -0500
  • 674ed88d13 * 2023/02/16 (v20.7.3) - Noise offset is recorded to the metadata. Thanks to space-nuko! - Show the moving average loss to prevent loss jumping in train_network.py and train_db.py. Thanks to shirayu! bmaltais 2023-02-17 19:18:11 -0500
  • f9863e3950 add dadapation to other trainers bmaltais 2023-02-16 19:33:46 -0500
  • 655f885cf4 Add dadapation to other trainers bmaltais 2023-02-16 19:33:33 -0500
  • 641a168e55 Integrate new kohya sd-script bmaltais 2023-02-14 18:52:08 -0500
  • a1f6438f7b Upgrade upgrade.ps1 script to fix reported issue: https://github.com/bmaltais/kohya_ss/issues/165 bmaltais 2023-02-14 17:42:36 -0500
  • 6129c7dd40 1st implementation bmaltais 2023-02-13 21:20:09 -0500
  • 261b6790ee Update tool bmaltais 2023-02-12 07:02:05 -0500
  • a008c62893
    Merge pull request #147 from bmaltais/dev bmaltais 2023-02-11 12:00:17 -0500
  • a49fb9cb8c 2023/02/11 (v20.7.2): - `lora_interrogator.py is added in networks folder. See python networks\lora_interrogator.py -h for usage. - For LoRAs where the activation word is unknown, this script compares the output of Text Encoder after applying LoRA to that of unapplied to find out which token is affected by LoRA. Hopefully you can figure out the activation word. LoRA trained with captions does not seem to be able to interrogate. - Batch size can be large (like 64 or 128). - train_textual_inversion.py` now supports multiple init words. - Following feature is reverted to be the same as before. Sorry for confusion: > Now the number of data in each batch is limited to the number of actual images (not duplicated). Because a certain bucket may contain smaller number of actual images, so the batch may contain same (duplicated) images. - Add new tool to sort, group and average crop image in a dataset bmaltais 2023-02-11 11:59:38 -0500
  • e5f8ba559f Add server_port and inbrowser support - to all gui scripts bmaltais 2023-02-10 08:22:03 -0500
  • dff9710c81
    Merge pull request #139 from bmaltais/dev bmaltais 2023-02-09 20:25:48 -0500
  • 56d171c55b Reverting changes to startup commands bmaltais 2023-02-09 20:25:42 -0500
  • 9660d83612
    Merge pull request #129 from jonathanzhang53/master bmaltais 2023-02-09 19:19:01 -0500
  • 9a7bb4c624
    Merge pull request #138 from bmaltais/dev bmaltais 2023-02-09 19:18:08 -0500
  • 7bc93821a0 2023/02/09 (v20.7.1) - Caption dropout is supported in `train_db.py, fine_tune.py and train_network.py. Thanks to forestsource! - --caption_dropout_rate option specifies the dropout rate for captions (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the image is trained with the empty caption. Default is 0 (no dropout). - --caption_dropout_every_n_epochs option specifies how many epochs to drop captions. If 3 is specified, in epoch 3, 6, 9 ..., images are trained with all captions empty. Default is None (no dropout). - --caption_tag_dropout_rate option specified the dropout rate for tags (comma separated tokens) (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the tag is removed from the caption. If --keep_tokens option is set, these tokens (tags) are not dropped. Default is 0 (no droupout). - The bulk image downsampling script is added. Documentation is [here](https://github.com/kohya-ss/sd-scripts/blob/main/train_network_README-ja.md#%E7%94%BB%E5%83%8F%E3%83%AA%E3%82%B5%E3%82%A4%E3%82%BA%E3%82%B9%E3%82%AF%E3%83%AA%E3%83%97%E3%83%88) (in Jpanaese). Thanks to bmaltais! - Typo check is added. Thanks to shirayu! - Add option to autolaunch the GUI in a browser and set the server_port. USe either gui.ps1 --inbrowser --server_port 3456or gui.cmd -inbrowser -server_port 3456` bmaltais 2023-02-09 19:17:24 -0500
  • 90c0d55457 2023/02/09 (v20.7.1) - Caption dropout is supported in `train_db.py, fine_tune.py and train_network.py. Thanks to forestsource! - --caption_dropout_rate option specifies the dropout rate for captions (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the image is trained with the empty caption. Default is 0 (no dropout). - --caption_dropout_every_n_epochs option specifies how many epochs to drop captions. If 3 is specified, in epoch 3, 6, 9 ..., images are trained with all captions empty. Default is None (no dropout). - --caption_tag_dropout_rate option specified the dropout rate for tags (comma separated tokens) (0~1.0, 0.1 means 10% chance for dropout). If dropout occurs, the tag is removed from the caption. If --keep_tokens option is set, these tokens (tags) are not dropped. Default is 0 (no droupout). - The bulk image downsampling script is added. Documentation is [here](https://github.com/kohya-ss/sd-scripts/blob/main/train_network_README-ja.md#%E7%94%BB%E5%83%8F%E3%83%AA%E3%82%B5%E3%82%A4%E3%82%BA%E3%82%B9%E3%82%AF%E3%83%AA%E3%83%97%E3%83%88) (in Jpanaese). Thanks to bmaltais! - Typo check is added. Thanks to shirayu! - Add option to autolaunch the GUI in a browser and set the server_port. USe either gui.ps1 --inbrowser --server_port 3456or gui.cmd -inbrowser -server_port 3456` bmaltais 2023-02-09 19:17:17 -0500
  • 6c4348233f README documentation update jonathanzhang53 2023-02-07 22:30:25 -0500
  • 09d3a72cd8 Adding support for caption dropout bmaltais 2023-02-07 20:58:35 -0500
  • 9d53e7bd80
    Merge pull request #118 from bmaltais/dev bmaltais 2023-02-06 11:04:55 -0500
  • 8d559ded18 * 2023/02/06 (v20.7.0) - `--bucket_reso_steps and --bucket_no_upscale options are added to training scripts (fine tuning, DreamBooth, LoRA and Textual Inversion) and prepare_buckets_latents.py. - --bucket_reso_steps takes the steps for buckets in aspect ratio bucketing. Default is 64, same as before. - Any value greater than or equal to 1 can be specified; 64 is highly recommended and a value divisible by 8 is recommended. - If less than 64 is specified, padding will occur within U-Net. The result is unknown. - If you specify a value that is not divisible by 8, it will be truncated to divisible by 8 inside VAE, because the size of the latent is 1/8 of the image size. - If --bucket_no_upscale option is specified, images smaller than the bucket size will be processed without upscaling. - Internally, a bucket smaller than the image size is created (for example, if the image is 300x300 and bucket_reso_steps=64, the bucket is 256x256). The image will be trimmed. - Implementation of [#130](https://github.com/kohya-ss/sd-scripts/issues/130). - Images with an area larger than the maximum size specified by --resolution are downsampled to the max bucket size. - Now the number of data in each batch is limited to the number of actual images (not duplicated). Because a certain bucket may contain smaller number of actual images, so the batch may contain same (duplicated) images. - --random_crop` now also works with buckets enabled. - Instead of always cropping the center of the image, the image is shifted left, right, up, and down to be used as the training data. This is expected to train to the edges of the image. - Implementation of discussion [#34](https://github.com/kohya-ss/sd-scripts/discussions/34). bmaltais 2023-02-06 11:04:07 -0500
  • cbfc311687 Integrate new bucket parameters in GUI bmaltais 2023-02-05 20:07:00 -0500
  • 2486af9903 Update to latest dev code of kohya_s. WIP bmaltais 2023-02-05 14:16:53 -0500
  • 18c551e579
    Merge pull request #107 from bmaltais/dev bmaltais 2023-02-04 11:55:32 -0500
  • 2626214f8a Add support for LoRA resizing bmaltais 2023-02-04 11:55:06 -0500
  • 2ed93b7a4f
    Merge pull request #105 from bmaltais/dev bmaltais 2023-02-04 08:37:25 -0500
  • 045750b46a v20.6.0 - Increase max LoRA rank (dim) size to 1024. - Update finetune preprocessing scripts. - `.bmp and .jpeg are supported. Thanks to breakcore2 and p1atdev! - The default weights of tag_images_by_wd14_tagger.py is now SmilingWolf/wd-v1-4-convnext-tagger-v2. You can specify another model id from SmilingWolf by --repo_id option. Thanks to SmilingWolf for the great work. - To change the weight, remove wd14_tagger_model folder, and run the script again. - --max_data_loader_n_workers option is added to each script. This option uses the DataLoader for data loading to speed up loading, 20%~30% faster. - Please specify 2 or 4, depends on the number of CPU cores. - --recursive option is added to merge_dd_tags_to_metadata.py and merge_captions_to_metadata.py, only works with --full_path. - make_captions_by_git.py is added. It uses [GIT microsoft/git-large-textcaps](https://huggingface.co/microsoft/git-large-textcaps) for captioning. - requirements.txt is updated. If you use this script, [please update the libraries](https://github.com/kohya-ss/sd-scripts#upgrade). - Usage is almost the same as make_captions.py, but batch size should be smaller. - --remove_words option removes as much text as possible (such as the word "XXXX" on it). - --skip_existing option is added to prepare_buckets_latents.py. Images with existing npz files are ignored by this option. - clean_captions_and_tags.py is updated to remove duplicated or conflicting tags, e.g. shirt is removed when white shirt exists. if black hair is with red hair, both are removed. - Tag frequency is added to the metadata in train_network.py. Thanks to space-nuko! - __All tags and number of occurrences of the tag are recorded.__ If you do not want it, disable metadata storing with --no_metadata` option. bmaltais 2023-02-04 08:36:35 -0500
  • 20e62af1a6 Update to latest kohya_ss sd-script code bmaltais 2023-02-03 14:40:03 -0500
  • 9c5bdd1749
    Merge pull request #83 from bmaltais/dev bmaltais 2023-01-30 08:27:00 -0500
  • c8f4c9d6e8 Add support for lr_scheduler_num_cycles, lr_scheduler_power bmaltais 2023-01-30 08:26:15 -0500
  • 2ec7432440 Fix issue 81: https://github.com/bmaltais/kohya_ss/issues/81 bmaltais 2023-01-29 11:17:30 -0500
  • d45a7abb46 Add reference to Linux docker port bmaltais 2023-01-29 11:12:05 -0500
  • bc8a4757f8 Sync with kohya 2023/01/29 update bmaltais 2023-01-29 11:10:06 -0500
  • fdd1b02a26
    Merge pull request #78 from bmaltais/dev bmaltais 2023-01-27 19:46:55 -0500
  • a4957cfea7 Adding LoRA tutorial bmaltais 2023-01-27 19:46:13 -0500
  • 202923b3ce Add support for --keep_token option bmaltais 2023-01-27 07:33:44 -0500
  • bf371b49bf Fix issue 71 bmaltais 2023-01-27 07:04:35 -0500
  • c6f87f1c01 Create save dir if not exist bmaltais 2023-01-26 21:05:36 -0500
  • 08d189400d
    Merge pull request #74 from bmaltais/dev bmaltais 2023-01-26 16:23:46 -0500
  • 03bd2e9b01 Add TI training support bmaltais 2023-01-26 16:22:58 -0500
  • 49bada0d25 Update default save model file format to safetensors bmaltais 2023-01-22 18:21:09 -0500
  • 511361c80b - Add new tool to verify LoRA weights produced by the trainer. Can be found under "Dreambooth LoRA/Tools/Verify LoRA bmaltais 2023-01-22 11:40:14 -0500