Replies: 1 comment 3 replies
-
|
I think pruning layers from Chroma might work better than pruning modulation out of a flux lite model. And the reason I think that is because Chroma training took a lot longer than Flux Lite's. It might be worth trying what you did in Flex.1 alpha instead of Flux.1 Lite. Both are 8B, but I'm pretty sure Flex.1 is based on Flux Schnell. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment


Uh oh!
There was an error while loading. Please reload this page.
-
So I tried transplanting the distilled guidance from chroma (I went for the first available model Chroma-v2.5) to freepiks flux.1-lite-8B. I also hacked some stuff in sd.cpp to make it work somewhat, but it only works a little.
Reasons I can think of that could cause this:
splicing details
Also needs sd.cpp changes to compensate it not removing the middle blocks modulation guidance.
What really should be done what lodestone did for chroma, but the code was never published I think. ( https://huggingface.co/lodestones/Chroma/discussions/12 )
Or the reverse can be done too, where we prune some middle layers from chroma. flux.1-lite-8B removed double blocks 5-15 and retrained 4 (or all later?).
Which would also be nicer license and feature wise.
Beta Was this translation helpful? Give feedback.
All reactions