mirror of
https://github.com/modelscope/DiffSynth-Studio.git
synced 2026-03-24 18:28:10 +00:00
bugfix
This commit is contained in:
@@ -117,6 +117,4 @@ if __name__ == "__main__":
|
|||||||
remove_prefix_in_ckpt=args.remove_prefix_in_ckpt,
|
remove_prefix_in_ckpt=args.remove_prefix_in_ckpt,
|
||||||
state_dict_converter=FluxLoRAConverter.align_to_opensource_format if args.align_to_opensource_format else lambda x:x,
|
state_dict_converter=FluxLoRAConverter.align_to_opensource_format if args.align_to_opensource_format else lambda x:x,
|
||||||
)
|
)
|
||||||
optimizer = torch.optim.AdamW(model.trainable_modules(), lr=args.learning_rate, weight_decay=args.weight_decay)
|
|
||||||
scheduler = torch.optim.lr_scheduler.ConstantLR(optimizer)
|
|
||||||
launch_training_task(dataset, model, model_logger, args=args)
|
launch_training_task(dataset, model, model_logger, args=args)
|
||||||
|
|||||||
@@ -126,6 +126,4 @@ if __name__ == "__main__":
|
|||||||
args.output_path,
|
args.output_path,
|
||||||
remove_prefix_in_ckpt=args.remove_prefix_in_ckpt
|
remove_prefix_in_ckpt=args.remove_prefix_in_ckpt
|
||||||
)
|
)
|
||||||
optimizer = torch.optim.AdamW(model.trainable_modules(), lr=args.learning_rate, weight_decay=args.weight_decay)
|
|
||||||
scheduler = torch.optim.lr_scheduler.ConstantLR(optimizer)
|
|
||||||
launch_training_task(dataset, model, model_logger, args=args)
|
launch_training_task(dataset, model, model_logger, args=args)
|
||||||
|
|||||||
Reference in New Issue
Block a user