-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implementation details #1
Comments
We use the same set of hyperparameters used in the original papers for each model. |
The dataset size is changed (for example CUB-Color should have fewer training images than the original CUB). Have you also adapted the training epoch number? |
I also encountered a problem when I was trying to reproduce DAMSM R-precision (for C-CUB-Color). In the original AttnGAN, there are 5450 tokens. I downloaded the pretrained text encoder provided by you. There are also 5450 token embeddings. However, I cannot tokenize the C-CUB-Color captions. For example the caption "this bird is black in color, with a black beak.", "." is not included in the word dictionary of the DAMSM module. How did you process this case? |
For test_unseen split i got an error when generating with the image_id 191.Red_headed_Woodpecker/Red_Headed_Woodpecker_0018_183455 It is too long and CLIP doesn't support this. What should I do for this? |
dear authors,
It's a great work:) Well, i would like to ask something about the implementation details.
I see DMGAN, ControlGAN and DFGAN are retrained on the new data split. These models are trained with different settings in their original papers (e.g. learning rate, epochs...). How do you select the training settings/hyperparameters when you retrained them:)
The text was updated successfully, but these errors were encountered: