Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Not normalizing the expert states when computing gradient penalty. #27

Open
EGalahad opened this issue Sep 27, 2024 · 3 comments
Open

Comments

@EGalahad
Copy link

grad_pen_loss = self.discriminator.compute_grad_pen(

When using state normalization, the sample_amp_expert tuple is not normalized.

@EGalahad
Copy link
Author

if self.amp_normalizer is not None:

and I think maybe we should update the normalizer stats before normalizing them?

@apirrone
Copy link

apirrone commented Oct 4, 2024

@EGalahad What effect do you think this has ?

@EGalahad
Copy link
Author

I just think it is weird to pass in the normalized values to update the normalizer. I think you should pass un-normalized version so that the normalizer can update running mean?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants