-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The model "loses" or "degrades" its multimodality when you increase the size of the model #33
Comments
@F4k3r22 turns out i introduced a bug when adding the meta tokens from hymba paper would you like to try again? |
Ok, I'm going to do some training tests, and then I'll let you know if it still has multimodality |
Size test:
Large model:
And if it helps at all I'll leave the training file in case there's an error in that:
Large model:
|
Hey, I'm going to do a test by increasing the dimension of the autoencoder along with the model, maybe it's because of this discrepancy in dimensionality that causes this error. |
I already did the test and it still seems to lose that multimodality, even though the auto encoder and Transfusion share the same dimensionality:
AutoEncoder:
|
The model "loses" or "degrades" its multimodality when you increase the size of the model, the maximum that can be increased is this:
I noticed this because I wanted to train with this setup and I could never get multimodality in sampling during my training:
Maybe it's because of the
transfusion_attn_mask
, we can give more weight to the multimodal interactions of the model to compensate for the larger possible configurationsI think this could be done:
Even if it's not the correct solution, I hope this observation helps you, have a nice day
The text was updated successfully, but these errors were encountered: