-
Notifications
You must be signed in to change notification settings - Fork 85
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
structure not being retained #27
Comments
UPDATE: i also ran the above test on 25 million iter model, in that result the model is more clearly and exactly recreating the structure images, a lot better than the 50 million trained model attaching the zip file for analysis. Could you please tell what could be the reason that the 50 mill model is recreating the structure poorly as compared to 25mill model https://drive.google.com/file/d/1PszJfCIoKxRKFAS4zEEv9td5hVTgD2VG/view?usp=sharing |
|
Hi @tom99763, I suppose it's because the patch discriminator becomes stronger during the course of training, encouraging the generator to make more changes. You can try the following two things:
|
Another question, if there's an auxiliary classifier in the discriminator, for example, classify facial expressions,can the strcture be retained(reconstruction) if using less downsample and smaller patches? Seems like the biggest contribution is the patch discriminator in this paper, have you ever try constrative loss or triplet loss in the discriminator? It maybe has the potential to extend to the combination of hard recognition task and generation task |
Having an auxiliary classifier is a good idea. I think you are right. I did think about contrastive loss or triplet loss, and that was our long term plan in case the current formulation does not work out. The concern was that it will be quite memory intensive, because you will need to encode images, swap the codes, decode them, and the re-encode to compute the contrastive loss. Fortunately the current formulation seemed to have enough inductive bias to generate interesting outputs. |
hi @taesungp great piece of work, I trained it on my dataset of 50k images for 50 Mil iterations as you suggested, on testing time the results are quite impressive but in some cases, the structure is not being correctly reconstructed, I would like that the shapes be generated almost the same(full swapping). What can be the problem will training further help?
![image](https://user-images.githubusercontent.com/55244143/159499184-c0a8bc24-dca3-4217-a73a-7f8e847c467f.png)
The text was updated successfully, but these errors were encountered: