-
Notifications
You must be signed in to change notification settings - Fork 268
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Detail architecture of dynamic instance normalization #18
Comments
Hi @zy-xc Thank you for your interest in our work. Here is the link for the corresponding supplement: https://drive.google.com/file/d/1sBFXqWaWOeMuaaVHMM-ddBssKr3OmutW/view?usp=sharing Please feel free to contact me if there is any other question. Thank you! Best, |
Thank you for your reply! I am a bit confusing about size of weight generated by Weight/Bias network. Is the dynamic convolution layer set groups = 64(num_channels of content feature) ? It seens that the size of style image should be large if we set groups=1. For example, considering standard DIN with kernel_size=1. The weight size generated by Weight Net should be 64 * 64 * 1 * 1. So the vgg features size of style image should be at least 64 * 64 * 64(C * H * W), and the size of style image should be at least 512 * 512. Then if we want to train standard DIN with kernel_size=3, the size of style image should be at least 1536 * 1536. Or standard DIN set groups=64 and the size of generated weight should be 64 * kernel_size * kernel_size ? Thank you! |
Hi @zy-xc Thank you for your interests in our work! Regarding your question, yes, we indeed set group # to be equal to the feature channel, which is indicated in the "Architecture Details" in the supplement. Also, please kindly note that the size of the generated weight and bias is not correlated with the input size, since we use an adaptive pooling layer in the corresponding weight and bias networks. You can set the desired size of the weight and bias by controlling the adaptive pooling layer. Please let me know if there is any other question. Thank you. Best, |
I find the supplementary detail confusing to implement. Has anyone implemented in Pytorch yet? Can you help me? |
Thank you for your interests in our work! Could you please elaborate which part exactly is confusing? I am more than happy to clarify it. Also, if you would like our source code, please drop me an email to apply for the necessary permission that is required by the company. Thanks! Best, |
Hi @ycjing |
Hi @ycjing
2.Res layer and upsampling layer is quite lacking in information and I don't know where it is on the illustration |
Thanks for your interests again! Please feel free to reach me if there is anything else that is not clear. Cheers, |
Could you please provide the detailed log information? Thanks! Best, |
Here is my test case:
Logs:
|
As depicted in the log, the group # is wrong, which should be equal to in_channel. Best, |
Hi @ycjing
|
Hi @ycjing
Here is my code:
Hope you help me. Thank you so much. |
|
Please refer to my previous reply and be careful about the output dimensions. Best, |
Hi @ycjing |
Thank you for your interests. From your code, I think you have already got the point, i.e., dynamically predicting the weight and bias via the weight and bias networks. Could you please further elaborate your question? Thanks! Best, |
Hi @ycjing |
No problem. The weight and bias are, actually, the output of the corresponding weight/bias networks, which is somewhat similar to the dynamic filter network (https://arxiv.org/abs/1605.09673). Cheers, |
Hi @ycjing |
Thank you for your interest. The values are, in fact, not the same. As demonstrated in the figure and explained in the paper, we use a separate weight net and bias net to produce the corresponding weight and bias. Best, |
Hello @ycjing
Thanks for your brilliant works! I am interesting in paper "Dynamic Instance Normalization for Arbitrary Style Transfer" but I don't know the detail architecture of DIN and can't find the supplementary material.
Would you please provide the detailed network architecture of this paper?
Thank you!
The text was updated successfully, but these errors were encountered: