Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Increasing token size #71

Open
jinnan-chen opened this issue Oct 31, 2024 · 3 comments
Open

Increasing token size #71

jinnan-chen opened this issue Oct 31, 2024 · 3 comments

Comments

@jinnan-chen
Copy link

Hi Tianhong,

I have trained the MAR on 1D unordered latents, it works fine for 256 tokens with 64 chanels, the loss converges at 0.35. However, when training on 1k or 2k tokens with 64 chanels, the loss converge at 0.45 and the results looks bad, even though the VAE reonstruction ability is higher than 256 tokens. Is there any suggestions? Thanks!

@LTH14
Copy link
Owner

LTH14 commented Oct 31, 2024

You might need to check the parameters such as vae_embed_dim, vae_stride, etc?

@jinnan-chen
Copy link
Author

Hi,
My tokens is not from 2D images, so I dont have vae_stride, and my token_embed_dim=vae_embed_dim=64.
When I use token_embed_dim=64, seq_len=buffer_size=256, it converges fast and generate good results.
So, when I increase the self.seq_len, should I increase the buffer_size during training and increase num_iter in sample_tokens accordingly?

@LTH14
Copy link
Owner

LTH14 commented Nov 2, 2024

buffer size does not need to be increased. num_iter should be increased (e.g., 128 for seq_len=1024)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants