You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the training of cyclegan-turbo, conv_in was set to be trainable, but it seems that the weights of conv_in after training were not saved. Instead, only the LoRA weights of conv_in were saved. Why is that?
Similarly, in pix2pix-turbo, conv_in was replaced with a TwinConv layer, and the paper does not seem to explain the function of this part.
Thanks
The text was updated successfully, but these errors were encountered:
I think the TwinConv layer is needed to sample diverse outputs as described on page 7 and 8 of the Paper in the section "Generating diverse outputs". If you only want to have a deterministic output you don't need it.
Your point that the conv_in is not saved might be valid, since I ran into the problem that my pretrained models did not behave well during inference. I did not check the issue but I ended up with saving all parameters and now it works - the downside of course is that my checkpoint files are very large.
In the training of cyclegan-turbo,
conv_in
was set to be trainable, but it seems that the weights ofconv_in
after training were not saved. Instead, only the LoRA weights ofconv_in
were saved. Why is that?Similarly, in pix2pix-turbo,
conv_in
was replaced with aTwinConv
layer, and the paper does not seem to explain the function of this part.Thanks
The text was updated successfully, but these errors were encountered: