Replies: 2 comments
-
From my years of lora training experience, dim = alpha yield the best result. Any alpha value less than dim will cause the learning rate reducing efficiency. To overcome the reducing LR efficiency, you need to increase your training steps. Too many steps will cause over training. |
Beta Was this translation helpful? Give feedback.
-
makes me think that less alpha == more dampening |
Beta Was this translation helpful? Give feedback.
-
Hello, could someone explain Network Alpha a bit? If there's an article that explains it clearly, please send it my way. I'm seeing a lot of conflicting information and am at this point not sure what to trust.
As I understand:
As I try to understand the article written here: https://github.com/bmaltais/kohya_ss/wiki/LoRA-training-parameters#network-alpha I am unclear which values are higher dampening and which are lower.
If [Network alpha] is set to the value of [Network dim], is it the most dampening possible or the least? (Or am I misunderstanding things entirely?)
Any clarification would be greatly appreciated.
Beta Was this translation helpful? Give feedback.
All reactions