Fix issue 278
This commit is contained in:
parent
d30abe5491
commit
29bb8599bb
@ -165,6 +165,7 @@ This will store your a backup file with your current locally installed pip packa
|
|||||||
|
|
||||||
* 2023/03/03 (v21.1.2):
|
* 2023/03/03 (v21.1.2):
|
||||||
- Fix issue https://github.com/bmaltais/kohya_ss/issues/277
|
- Fix issue https://github.com/bmaltais/kohya_ss/issues/277
|
||||||
|
- Fix issue https://github.com/bmaltais/kohya_ss/issues/278 introduce by LoCon project switching to pip module. Make sure to run upgrade.ps1 to install the latest pip requirements for LoCon support.
|
||||||
* 2023/03/02 (v21.1.1):
|
* 2023/03/02 (v21.1.1):
|
||||||
- Emergency fix for https://github.com/bmaltais/kohya_ss/issues/261
|
- Emergency fix for https://github.com/bmaltais/kohya_ss/issues/261
|
||||||
* 2023/03/02 (v21.1.0):
|
* 2023/03/02 (v21.1.0):
|
||||||
|
61
lora_gui.py
61
lora_gui.py
@ -48,18 +48,6 @@ save_style_symbol = '\U0001f4be' # 💾
|
|||||||
document_symbol = '\U0001F4C4' # 📄
|
document_symbol = '\U0001F4C4' # 📄
|
||||||
path_of_this_folder = os.getcwd()
|
path_of_this_folder = os.getcwd()
|
||||||
|
|
||||||
def getlocon(existance):
|
|
||||||
now_path = os.getcwd()
|
|
||||||
if existance:
|
|
||||||
print('Checking LoCon script version...')
|
|
||||||
os.chdir(os.path.join(path_of_this_folder, 'locon'))
|
|
||||||
os.system('git pull')
|
|
||||||
os.chdir(now_path)
|
|
||||||
else:
|
|
||||||
os.chdir(path_of_this_folder)
|
|
||||||
os.system('git clone https://github.com/KohakuBlueleaf/LoCon.git locon')
|
|
||||||
os.chdir(now_path)
|
|
||||||
|
|
||||||
|
|
||||||
def save_configuration(
|
def save_configuration(
|
||||||
save_as,
|
save_as,
|
||||||
@ -122,8 +110,11 @@ def save_configuration(
|
|||||||
caption_dropout_every_n_epochs,
|
caption_dropout_every_n_epochs,
|
||||||
caption_dropout_rate,
|
caption_dropout_rate,
|
||||||
optimizer,
|
optimizer,
|
||||||
optimizer_args,noise_offset,
|
optimizer_args,
|
||||||
LoRA_type='Standard', conv_dim=0, conv_alpha=0,
|
noise_offset,
|
||||||
|
LoRA_type='Standard',
|
||||||
|
conv_dim=0,
|
||||||
|
conv_alpha=0,
|
||||||
):
|
):
|
||||||
# Get list of function parameters and values
|
# Get list of function parameters and values
|
||||||
parameters = list(locals().items())
|
parameters = list(locals().items())
|
||||||
@ -230,8 +221,11 @@ def open_configuration(
|
|||||||
caption_dropout_every_n_epochs,
|
caption_dropout_every_n_epochs,
|
||||||
caption_dropout_rate,
|
caption_dropout_rate,
|
||||||
optimizer,
|
optimizer,
|
||||||
optimizer_args,noise_offset,
|
optimizer_args,
|
||||||
LoRA_type='Standard', conv_dim=0, conv_alpha=0,
|
noise_offset,
|
||||||
|
LoRA_type='Standard',
|
||||||
|
conv_dim=0,
|
||||||
|
conv_alpha=0,
|
||||||
):
|
):
|
||||||
# Get list of function parameters and values
|
# Get list of function parameters and values
|
||||||
parameters = list(locals().items())
|
parameters = list(locals().items())
|
||||||
@ -324,8 +318,11 @@ def train_model(
|
|||||||
caption_dropout_every_n_epochs,
|
caption_dropout_every_n_epochs,
|
||||||
caption_dropout_rate,
|
caption_dropout_rate,
|
||||||
optimizer,
|
optimizer,
|
||||||
optimizer_args,noise_offset,
|
optimizer_args,
|
||||||
LoRA_type, conv_dim, conv_alpha,
|
noise_offset,
|
||||||
|
LoRA_type,
|
||||||
|
conv_dim,
|
||||||
|
conv_alpha,
|
||||||
):
|
):
|
||||||
if pretrained_model_name_or_path == '':
|
if pretrained_model_name_or_path == '':
|
||||||
msgbox('Source model information is missing')
|
msgbox('Source model information is missing')
|
||||||
@ -462,9 +459,15 @@ def train_model(
|
|||||||
if not float(prior_loss_weight) == 1.0:
|
if not float(prior_loss_weight) == 1.0:
|
||||||
run_cmd += f' --prior_loss_weight={prior_loss_weight}'
|
run_cmd += f' --prior_loss_weight={prior_loss_weight}'
|
||||||
if LoRA_type == 'LoCon':
|
if LoRA_type == 'LoCon':
|
||||||
getlocon(os.path.exists(os.path.join(path_of_this_folder, 'locon')))
|
try:
|
||||||
run_cmd += f' --network_module=locon.locon.locon_kohya'
|
import locon.locon_kohya
|
||||||
run_cmd += f' --network_args "conv_dim={conv_dim}" "conv_alpha={conv_alpha}"'
|
except ModuleNotFoundError:
|
||||||
|
print("\033[1;31mError:\033[0m The required module 'locon' is not installed. Please install by running \033[33mupgrade.ps1\033[0m before running this program.")
|
||||||
|
return
|
||||||
|
run_cmd += f' --network_module=locon.locon_kohya'
|
||||||
|
run_cmd += (
|
||||||
|
f' --network_args "conv_dim={conv_dim}" "conv_alpha={conv_alpha}"'
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
run_cmd += f' --network_module=networks.lora'
|
run_cmd += f' --network_module=networks.lora'
|
||||||
|
|
||||||
@ -646,7 +649,7 @@ def lora_tab(
|
|||||||
'Standard',
|
'Standard',
|
||||||
'LoCon',
|
'LoCon',
|
||||||
],
|
],
|
||||||
value='Standard'
|
value='Standard',
|
||||||
)
|
)
|
||||||
lora_network_weights = gr.Textbox(
|
lora_network_weights = gr.Textbox(
|
||||||
label='LoRA network weights',
|
label='LoRA network weights',
|
||||||
@ -710,8 +713,9 @@ def lora_tab(
|
|||||||
)
|
)
|
||||||
|
|
||||||
with gr.Group(visible=False) as LoCon_group:
|
with gr.Group(visible=False) as LoCon_group:
|
||||||
|
|
||||||
def LoRA_type_change(LoRA_type):
|
def LoRA_type_change(LoRA_type):
|
||||||
if LoRA_type == "LoCon":
|
if LoRA_type == 'LoCon':
|
||||||
return gr.Group.update(visible=True)
|
return gr.Group.update(visible=True)
|
||||||
else:
|
else:
|
||||||
return gr.Group.update(visible=False)
|
return gr.Group.update(visible=False)
|
||||||
@ -734,7 +738,9 @@ def lora_tab(
|
|||||||
label='LoCon Convolution Alpha',
|
label='LoCon Convolution Alpha',
|
||||||
)
|
)
|
||||||
# Show of hide LoCon conv settings depending on LoRA type selection
|
# Show of hide LoCon conv settings depending on LoRA type selection
|
||||||
LoRA_type.change(LoRA_type_change, inputs=[LoRA_type], outputs=[LoCon_group])
|
LoRA_type.change(
|
||||||
|
LoRA_type_change, inputs=[LoRA_type], outputs=[LoCon_group]
|
||||||
|
)
|
||||||
with gr.Row():
|
with gr.Row():
|
||||||
max_resolution = gr.Textbox(
|
max_resolution = gr.Textbox(
|
||||||
label='Max resolution',
|
label='Max resolution',
|
||||||
@ -894,8 +900,11 @@ def lora_tab(
|
|||||||
caption_dropout_every_n_epochs,
|
caption_dropout_every_n_epochs,
|
||||||
caption_dropout_rate,
|
caption_dropout_rate,
|
||||||
optimizer,
|
optimizer,
|
||||||
optimizer_args,noise_offset,
|
optimizer_args,
|
||||||
LoRA_type, conv_dim, conv_alpha,
|
noise_offset,
|
||||||
|
LoRA_type,
|
||||||
|
conv_dim,
|
||||||
|
conv_alpha,
|
||||||
]
|
]
|
||||||
|
|
||||||
button_open_config.click(
|
button_open_config.click(
|
||||||
|
@ -19,10 +19,10 @@ transformers==4.26.0
|
|||||||
fairscale==0.4.13
|
fairscale==0.4.13
|
||||||
requests==2.28.2
|
requests==2.28.2
|
||||||
timm==0.6.12
|
timm==0.6.12
|
||||||
# for WD14 captioning
|
|
||||||
# tensorflow<2.11
|
# tensorflow<2.11
|
||||||
huggingface-hub==0.12.0
|
huggingface-hub==0.12.0
|
||||||
tensorflow==2.10.1
|
tensorflow==2.10.1
|
||||||
# xformers @ https://github.com/C43H66N12O12S2/stable-diffusion-webui/releases/download/f/xformers-0.0.14.dev0-cp310-cp310-win_amd64.whl
|
# For locon support
|
||||||
|
locon==0.0.2
|
||||||
# for kohya_ss library
|
# for kohya_ss library
|
||||||
.
|
.
|
Loading…
Reference in New Issue
Block a user