-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About the rich feedback model release #3
Comments
Hello, currently we don't have the plan to release the model. |
@leebird Looking forward to the rich feedback model... |
Looking forward |
Thanks for all the interests in our work! Due to company policies (related to productization etc.) we could not open source the model. We have included details of how to reproduce the results in our paper. If you have further questions please email the corresponding authors, and we'd be happy to help you reproduce the results. |
Hello, can you tell me how you trained the reward model, like which layers were frozen and which tuning method was used? |
Hi @srymaker , we finetuned all the layers in the model, including the ViT component. We tried freezing the ViT component but it didn't work well, especially for the heatmap tasks. Experiment details including hyperparameters and optimizer can be found in Section 9 in the paper. |
Thank you for your answer. Do all layers refer to the encoder and decoder in t5? |
@srymaker yes, all the layers are from the ViT and T5 encoder/decoder. Note there is a pretraining stage for the ViT and T5 layers on multimodal data as they were originally pretrained on unimodal data only. |
Hello, For single-modal pretraining tasks, are only the natural image captioning tasks on the WebLI dataset used? What other tasks are included? |
Thanks for your great work!
Will the rich feedback model be released?I'd love to test and experience the model and apply it to my own tasks!
The text was updated successfully, but these errors were encountered: