From ee0ee9468aebe6ff0e55ecba22150348dc22c6a2 Mon Sep 17 00:00:00 2001 From: Mary Hipp Date: Tue, 5 Aug 2025 16:11:16 -0400 Subject: [PATCH] update default for transformer_learning_rate --- src/invoke_training/pipelines/flux/lora/config.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/invoke_training/pipelines/flux/lora/config.py b/src/invoke_training/pipelines/flux/lora/config.py index e631fb71..eeaa4f7a 100644 --- a/src/invoke_training/pipelines/flux/lora/config.py +++ b/src/invoke_training/pipelines/flux/lora/config.py @@ -55,7 +55,7 @@ class FluxLoraConfig(BasePipelineConfig): rate. """ - transformer_learning_rate: float | None = 1e-4 + transformer_learning_rate: float | None = 4e-4 """The learning rate to use for the transformer model. If set, this overrides the optimizer's default learning rate. """