Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Dataset sizes #2

Open
bob80333 opened this issue Jun 23, 2019 · 1 comment
Open

Dataset sizes #2

bob80333 opened this issue Jun 23, 2019 · 1 comment

Comments

@bob80333
Copy link

bob80333 commented Jun 23, 2019

Do your datasets have to be the same size? Or do they just have to be divisible by batch size? I was trying this repo out but ran into some issues where randomly the losses would become NaN.

My trainA has 33344 images and my trainB has 22933 images. All hyperparams were default except dataset paths/names.

@WonwoongCho
Copy link
Owner

Hi bob80333,
The number of data between different domains (trainA and trainB) is not necessarily the same.

Actually, I'm not sure it is due to the same issue,
but empirically, I observed NaN during training when my dataset is too noisy to learn a translation. For example, if trainA data is vector illustrations, it rarely has consistent patterns across the dataset, thus it fails to learn. Therefore, I recommend you to carefully watch your dataset first in order to verify that a domain has a coherent pattern.

Please report again if the issue is not settled.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants