Fix issue 278
This commit is contained in:
parent
d30abe5491
commit
29bb8599bb
@ -165,6 +165,7 @@ This will store your a backup file with your current locally installed pip packa
|
||||
|
||||
* 2023/03/03 (v21.1.2):
|
||||
- Fix issue https://github.com/bmaltais/kohya_ss/issues/277
|
||||
- Fix issue https://github.com/bmaltais/kohya_ss/issues/278 introduce by LoCon project switching to pip module. Make sure to run upgrade.ps1 to install the latest pip requirements for LoCon support.
|
||||
* 2023/03/02 (v21.1.1):
|
||||
- Emergency fix for https://github.com/bmaltais/kohya_ss/issues/261
|
||||
* 2023/03/02 (v21.1.0):
|
||||
|
61
lora_gui.py
61
lora_gui.py
@ -48,18 +48,6 @@ save_style_symbol = '\U0001f4be' # 💾
|
||||
document_symbol = '\U0001F4C4' # 📄
|
||||
path_of_this_folder = os.getcwd()
|
||||
|
||||
def getlocon(existance):
|
||||
now_path = os.getcwd()
|
||||
if existance:
|
||||
print('Checking LoCon script version...')
|
||||
os.chdir(os.path.join(path_of_this_folder, 'locon'))
|
||||
os.system('git pull')
|
||||
os.chdir(now_path)
|
||||
else:
|
||||
os.chdir(path_of_this_folder)
|
||||
os.system('git clone https://github.com/KohakuBlueleaf/LoCon.git locon')
|
||||
os.chdir(now_path)
|
||||
|
||||
|
||||
def save_configuration(
|
||||
save_as,
|
||||
@ -122,8 +110,11 @@ def save_configuration(
|
||||
caption_dropout_every_n_epochs,
|
||||
caption_dropout_rate,
|
||||
optimizer,
|
||||
optimizer_args,noise_offset,
|
||||
LoRA_type='Standard', conv_dim=0, conv_alpha=0,
|
||||
optimizer_args,
|
||||
noise_offset,
|
||||
LoRA_type='Standard',
|
||||
conv_dim=0,
|
||||
conv_alpha=0,
|
||||
):
|
||||
# Get list of function parameters and values
|
||||
parameters = list(locals().items())
|
||||
@ -230,8 +221,11 @@ def open_configuration(
|
||||
caption_dropout_every_n_epochs,
|
||||
caption_dropout_rate,
|
||||
optimizer,
|
||||
optimizer_args,noise_offset,
|
||||
LoRA_type='Standard', conv_dim=0, conv_alpha=0,
|
||||
optimizer_args,
|
||||
noise_offset,
|
||||
LoRA_type='Standard',
|
||||
conv_dim=0,
|
||||
conv_alpha=0,
|
||||
):
|
||||
# Get list of function parameters and values
|
||||
parameters = list(locals().items())
|
||||
@ -324,8 +318,11 @@ def train_model(
|
||||
caption_dropout_every_n_epochs,
|
||||
caption_dropout_rate,
|
||||
optimizer,
|
||||
optimizer_args,noise_offset,
|
||||
LoRA_type, conv_dim, conv_alpha,
|
||||
optimizer_args,
|
||||
noise_offset,
|
||||
LoRA_type,
|
||||
conv_dim,
|
||||
conv_alpha,
|
||||
):
|
||||
if pretrained_model_name_or_path == '':
|
||||
msgbox('Source model information is missing')
|
||||
@ -462,9 +459,15 @@ def train_model(
|
||||
if not float(prior_loss_weight) == 1.0:
|
||||
run_cmd += f' --prior_loss_weight={prior_loss_weight}'
|
||||
if LoRA_type == 'LoCon':
|
||||
getlocon(os.path.exists(os.path.join(path_of_this_folder, 'locon')))
|
||||
run_cmd += f' --network_module=locon.locon.locon_kohya'
|
||||
run_cmd += f' --network_args "conv_dim={conv_dim}" "conv_alpha={conv_alpha}"'
|
||||
try:
|
||||
import locon.locon_kohya
|
||||
except ModuleNotFoundError:
|
||||
print("\033[1;31mError:\033[0m The required module 'locon' is not installed. Please install by running \033[33mupgrade.ps1\033[0m before running this program.")
|
||||
return
|
||||
run_cmd += f' --network_module=locon.locon_kohya'
|
||||
run_cmd += (
|
||||
f' --network_args "conv_dim={conv_dim}" "conv_alpha={conv_alpha}"'
|
||||
)
|
||||
else:
|
||||
run_cmd += f' --network_module=networks.lora'
|
||||
|
||||
@ -646,7 +649,7 @@ def lora_tab(
|
||||
'Standard',
|
||||
'LoCon',
|
||||
],
|
||||
value='Standard'
|
||||
value='Standard',
|
||||
)
|
||||
lora_network_weights = gr.Textbox(
|
||||
label='LoRA network weights',
|
||||
@ -710,8 +713,9 @@ def lora_tab(
|
||||
)
|
||||
|
||||
with gr.Group(visible=False) as LoCon_group:
|
||||
|
||||
def LoRA_type_change(LoRA_type):
|
||||
if LoRA_type == "LoCon":
|
||||
if LoRA_type == 'LoCon':
|
||||
return gr.Group.update(visible=True)
|
||||
else:
|
||||
return gr.Group.update(visible=False)
|
||||
@ -734,7 +738,9 @@ def lora_tab(
|
||||
label='LoCon Convolution Alpha',
|
||||
)
|
||||
# Show of hide LoCon conv settings depending on LoRA type selection
|
||||
LoRA_type.change(LoRA_type_change, inputs=[LoRA_type], outputs=[LoCon_group])
|
||||
LoRA_type.change(
|
||||
LoRA_type_change, inputs=[LoRA_type], outputs=[LoCon_group]
|
||||
)
|
||||
with gr.Row():
|
||||
max_resolution = gr.Textbox(
|
||||
label='Max resolution',
|
||||
@ -894,8 +900,11 @@ def lora_tab(
|
||||
caption_dropout_every_n_epochs,
|
||||
caption_dropout_rate,
|
||||
optimizer,
|
||||
optimizer_args,noise_offset,
|
||||
LoRA_type, conv_dim, conv_alpha,
|
||||
optimizer_args,
|
||||
noise_offset,
|
||||
LoRA_type,
|
||||
conv_dim,
|
||||
conv_alpha,
|
||||
]
|
||||
|
||||
button_open_config.click(
|
||||
|
@ -19,10 +19,10 @@ transformers==4.26.0
|
||||
fairscale==0.4.13
|
||||
requests==2.28.2
|
||||
timm==0.6.12
|
||||
# for WD14 captioning
|
||||
# tensorflow<2.11
|
||||
huggingface-hub==0.12.0
|
||||
tensorflow==2.10.1
|
||||
# xformers @ https://github.com/C43H66N12O12S2/stable-diffusion-webui/releases/download/f/xformers-0.0.14.dev0-cp310-cp310-win_amd64.whl
|
||||
# For locon support
|
||||
locon==0.0.2
|
||||
# for kohya_ss library
|
||||
.
|
Loading…
Reference in New Issue
Block a user