Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Different results of halfcheetah #1

Open
rainbow979 opened this issue Jan 17, 2023 · 7 comments
Open

Different results of halfcheetah #1

rainbow979 opened this issue Jan 17, 2023 · 7 comments

Comments

@rainbow979
Copy link

Hello,

I tried the same config with the repo and got the same good performance with the paper. However, when I tried the env halfcheetah and the testing score is much lower than the results from the paper. I only changed the returns_scale from 400 to 800 since halfcheetah had higher discounted returns. The training loss is shown in the figure. The blue line is halfcheetah and the purple line is hopper. I am wondering if there are other hyperparemeters needs to be changed.
1673933898018

@WangYong-Design
Copy link

Have you kept track of how long it takes to run an experiment on GPU?

@Looomo
Copy link

Looomo commented Jun 11, 2023

Hi, I was trying to reproduce the results of DD, but I couldnt make it, results on each dataset differ significantly from the results presented in the paper in Tab1. (I have tried only on 4 datasets, walker2d-medium-replay-v2, walker2d-medium-v2,halfcheetah-medium-replay-v2, and halfcheetah-medium-v2). So have you change anything in this repo or just directly run the downloaded code? Thanks!
Also, did you find the codes of length-K history condition in this repo?

@RenMing-Huang
Copy link

same question

1 similar comment
@SpaceLearner
Copy link

same question

@xishuxishu
Copy link

Hello,

I tried the same config with the repo and got the same good performance with the paper. However, when I tried the env halfcheetah and the testing score is much lower than the results from the paper. I only changed the returns_scale from 400 to 800 since halfcheetah had higher discounted returns. The training loss is shown in the figure. The blue line is halfcheetah and the purple line is hopper. I am wondering if there are other hyperparemeters needs to be changed. 1673933898018

hello,Have you solved this problem

@wangerlie
Copy link

wangerlie commented Nov 12, 2024

Hello,

I tried the same config with the repo and got the same good performance with the paper. However, when I tried the env halfcheetah and the testing score is much lower than the results from the paper. I only changed the returns_scale from 400 to 800 since halfcheetah had higher discounted returns. The training loss is shown in the figure. The blue line is halfcheetah and the purple line is hopper. I am wondering if there are other hyperparemeters needs to be changed. 1673933898018

hello @rainbow979
I also tried to reproduce the results, I run the code directly without ANY change in the code, I just couldn't get the same result in 'hopper-medium-expert-v2'. I am wondering how to reproduce the result in 'hopper-medium-expert-v2', have you ever change any hyperparameters? Thanks a lot .

@wangerlie
Copy link

I am also wondering do you have the same warning as following, I think it's not reasonable, there may have some problem within the normalization code?
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants