706dfe157f
* Merge both dreambooth and finetune back in one repo
153 lines
6.1 KiB
PowerShell
153 lines
6.1 KiB
PowerShell
# variables related to the pretrained model
|
|
$pretrained_model_name_or_path = "D:\models\test\samdoesart2\model\last"
|
|
$v2 = 1 # set to 1 for true or 0 for false
|
|
$v_model = 0 # set to 1 for true or 0 for false
|
|
|
|
# variables related to the training dataset and output directory
|
|
$train_dir = "D:\models\test\samdoesart2"
|
|
$image_folder = "D:\dataset\samdoesart2\raw"
|
|
$output_dir = "D:\models\test\samdoesart2\model_e2\"
|
|
$max_resolution = "512,512"
|
|
|
|
# variables related to the training process
|
|
$learning_rate = 1e-6
|
|
$lr_scheduler = "constant" # Default is constant
|
|
$lr_warmup = 0 # % of steps to warmup for 0 - 100. Default is 0.
|
|
$dataset_repeats = 40
|
|
$train_batch_size = 8
|
|
$epoch = 1
|
|
$save_every_n_epochs = 1
|
|
$mixed_precision = "bf16"
|
|
$save_precision = "fp16" # use fp16 for better compatibility with auto1111 and other repo
|
|
$seed = "494481440"
|
|
$num_cpu_threads_per_process = 6
|
|
$train_text_encoder = 0 # set to 1 to train text encoder otherwise set to 0
|
|
|
|
# variables related to the resulting diffuser model. If input is ckpt or tensors then it is not applicable
|
|
$convert_to_safetensors = 1 # set to 1 to convert resulting diffuser to ckpt
|
|
$convert_to_ckpt = 1 # set to 1 to convert resulting diffuser to ckpt
|
|
|
|
# other variables
|
|
$kohya_finetune_repo_path = "D:\kohya_ss"
|
|
|
|
### You should not need to change things below
|
|
|
|
# Set variables to useful values using ternary operator
|
|
$v_model = ($v_model -eq 0) ? $null : "--v_parameterization"
|
|
$v2 = ($v2 -eq 0) ? $null : "--v2"
|
|
$train_text_encoder = ($train_text_encoder -eq 0) ? $null : "--train_text_encoder"
|
|
|
|
# stop script on error
|
|
$ErrorActionPreference = "Stop"
|
|
|
|
# define a list of substrings to search for
|
|
$substrings_v2 = "stable-diffusion-2-1-base", "stable-diffusion-2-base"
|
|
|
|
# check if $v2 and $v_model are empty and if $pretrained_model_name_or_path contains any of the substrings in the v2 list
|
|
if ($v2 -eq $null -and $v_model -eq $null -and ($substrings_v2 | Where-Object { $pretrained_model_name_or_path -match $_ }).Count -gt 0) {
|
|
Write-Host("SD v2 model detected. Setting --v2 parameter")
|
|
$v2 = "--v2"
|
|
$v_model = $null
|
|
}
|
|
|
|
# define a list of substrings to search for v-objective
|
|
$substrings_v_model = "stable-diffusion-2-1", "stable-diffusion-2"
|
|
|
|
# check if $v2 and $v_model are empty and if $pretrained_model_name_or_path contains any of the substrings in the v_model list
|
|
elseif ($v2 -eq $null -and $v_model -eq $null -and ($substrings_v_model | Where-Object { $pretrained_model_name_or_path -match $_ }).Count -gt 0) {
|
|
Write-Host("SD v2 v_model detected. Setting --v2 parameter and --v_parameterization")
|
|
$v2 = "--v2"
|
|
$v_model = "--v_parameterization"
|
|
}
|
|
|
|
# activate venv
|
|
cd $kohya_finetune_repo_path
|
|
.\venv\Scripts\activate
|
|
|
|
# create caption json file
|
|
if (!(Test-Path -Path $train_dir)) {
|
|
New-Item -Path $train_dir -ItemType "directory"
|
|
}
|
|
|
|
python $kohya_finetune_repo_path\script\merge_captions_to_metadata.py `
|
|
--caption_extention ".txt" $image_folder $train_dir"\meta_cap.json"
|
|
|
|
# create images buckets
|
|
python $kohya_finetune_repo_path\script\prepare_buckets_latents.py `
|
|
$image_folder `
|
|
$train_dir"\meta_cap.json" `
|
|
$train_dir"\meta_lat.json" `
|
|
$pretrained_model_name_or_path `
|
|
--batch_size 4 --max_resolution $max_resolution --mixed_precision $mixed_precision
|
|
|
|
# Get number of valid images
|
|
$image_num = Get-ChildItem "$image_folder" -Recurse -File -Include *.npz | Measure-Object | % { $_.Count }
|
|
|
|
$repeats = $image_num * $dataset_repeats
|
|
Write-Host("Repeats = $repeats")
|
|
|
|
# calculate max_train_set
|
|
$max_train_set = [Math]::Ceiling($repeats / $train_batch_size * $epoch)
|
|
Write-Host("max_train_set = $max_train_set")
|
|
|
|
$lr_warmup_steps = [Math]::Round($lr_warmup * $max_train_set / 100)
|
|
Write-Host("lr_warmup_steps = $lr_warmup_steps")
|
|
|
|
Write-Host("$v2 $v_model")
|
|
|
|
accelerate launch --num_cpu_threads_per_process $num_cpu_threads_per_process $kohya_finetune_repo_path\script\fine_tune.py `
|
|
$v2 `
|
|
$v_model `
|
|
--pretrained_model_name_or_path=$pretrained_model_name_or_path `
|
|
--in_json $train_dir\meta_lat.json `
|
|
--train_data_dir="$image_folder" `
|
|
--output_dir=$output_dir `
|
|
--train_batch_size=$train_batch_size `
|
|
--dataset_repeats=$dataset_repeats `
|
|
--learning_rate=$learning_rate `
|
|
--lr_scheduler=$lr_scheduler `
|
|
--lr_warmup_steps=$lr_warmup_steps `
|
|
--max_train_steps=$max_train_set `
|
|
--use_8bit_adam `
|
|
--xformers `
|
|
--mixed_precision=$mixed_precision `
|
|
--save_every_n_epochs=$save_every_n_epochs `
|
|
--seed=$seed `
|
|
$train_text_encoder `
|
|
--save_precision=$save_precision
|
|
|
|
# check if $output_dir\last is a directory... therefore it is a diffuser model
|
|
if (Test-Path "$output_dir\last" -PathType Container) {
|
|
if ($convert_to_ckpt) {
|
|
Write-Host("Converting diffuser model $output_dir\last to $output_dir\last.ckpt")
|
|
python "$kohya_finetune_repo_path\tools\convert_diffusers20_original_sd.py" `
|
|
$output_dir\last `
|
|
$output_dir\last.ckpt `
|
|
--$save_precision
|
|
}
|
|
if ($convert_to_safetensors) {
|
|
Write-Host("Converting diffuser model $output_dir\last to $output_dir\last.safetensors")
|
|
python "$kohya_finetune_repo_path\tools\convert_diffusers20_original_sd.py" `
|
|
$output_dir\last `
|
|
$output_dir\last.safetensors `
|
|
--$save_precision
|
|
}
|
|
}
|
|
|
|
# define a list of substrings to search for inference file
|
|
$substrings_sd_model = ".ckpt", ".safetensors"
|
|
$matching_extension = foreach ($ext in $substrings_sd_model) {
|
|
Get-ChildItem $output_dir -File | Where-Object { $_.Extension -contains $ext }
|
|
}
|
|
|
|
if ($matching_extension.Count -gt 0) {
|
|
# copy the file named "v2-inference.yaml" from the "v2_inference" folder to $output_dir as last.yaml
|
|
if ( $v2 -ne $null -and $v_model -ne $null) {
|
|
Write-Host("Saving v2-inference-v.yaml as $output_dir\last.yaml")
|
|
Copy-Item -Path "$kohya_finetune_repo_path\v2_inference\v2-inference-v.yaml" -Destination "$output_dir\last.yaml"
|
|
}
|
|
elseif ( $v2 -ne $null ) {
|
|
Write-Host("Saving v2-inference.yaml as $output_dir\last.yaml")
|
|
Copy-Item -Path "$kohya_finetune_repo_path\v2_inference\v2-inference.yaml" -Destination "$output_dir\last.yaml"
|
|
}
|
|
} |