Which cross-attention optimization technique is best? Could someone please clarify when to use it and why? #846
Replies: 2 comments 4 replies
-
simple answer - it depends :) first was split-attention, then came invokeai, then doggetx. those 3 are somewhat platform agnostic, so they can be used regardless. xformers were heavily optimized for torch+cuda on nvidia, so if you have that, they used to be the best. and then second line is optional - since cross-optimization can lead to non-derministic results (tiny differences in images even if settings are the same) - this enables deterministic mode as some small cost. if there was a simple answer, it would be pre-set and there wouldn't be a configurable option. |
Beta Was this translation helpful? Give feedback.
-
On a RX 6900 XT I find that sub quadratic is the slowest, but uses the least amount of vram. This is useful when you want to generate something high resolution without getting OOM issues. In my specific case it would actually be great if it could switch between SDP and sub quadratic depending on predicted memory usage or simply just above some certain resolution. |
Beta Was this translation helpful? Give feedback.
-
simple question, can someone explain me please?
Beta Was this translation helpful? Give feedback.
All reactions