Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Submission for Issue #120 #151

Open
wants to merge 5 commits into
base: master
Choose a base branch
from

Conversation

yashkant
Copy link

@yashkant yashkant commented Jan 7, 2019

#120

Participant Information

  • Harshal Mittal, Indian Institute of Technology, Roorkee
  • Kartikey Pandey, Indian Institute of Technology, Roorkee
  • Yash Kant, Indian Institute of Technology, Roorkee

@reproducibility-org reproducibility-org added the checks-complete Submission criteria checks complete label Jan 8, 2019
@koustuvsinha koustuvsinha added reviewer-assigned Reviewer has been assigned and removed reviewer-assigned Reviewer has been assigned labels Feb 1, 2019
@reproducibility-org
Copy link
Collaborator

Hi, please find below a review submitted by one of the reviewers:

Score: 8
Reviewer 1 comment : This work tries to reproduce Partially Adam on Cifar100 and Cifar10 with VGG, (Wide-)ResNet. The experiments show that PAdam achieves better results. Discussion on future study is appreciated. The writing of this paper can be improved. For example, the figure of network architecture in page 4 is not necessary since it is not relevant to main topic (optimizer). Figure 2 is ill-formatted.
Confidence : 3

@reproducibility-org
Copy link
Collaborator

Hi, please find below a review submitted by one of the reviewers:

Score: 4
Reviewer 3 comment : The paper provides a good summary of the main contributions of the original work. The goal is to do a direction reproduction of the tables and graphs in the paper, using independent code. The proposed method is quite simple to implement, and the reproducibility builds on a common implementation of Adam as a starting point. The authors of the reproducibility study do not report any difficulty in implementing PADAM.

The reproducibility study considers the same hyperparameters as the original work for most of the results presented. There is a sub-section that considers interaction between a key parameter of PADAM (“p”) and the learning rate, which is novel. Unfortunately results for this were not finished and the plots only show a small number of training iterations; not enough to draw useful conclusions. In the other experiments, the reproducibility study follows directly from the original paper. I was surprised to see that the reproducibility study did not comment more on some differences. For example, the test error in Fig.2 of the reproducibility study seems to have much more variance than corresponding plots in Fig.2 of the original work. Do you agree with this observation? Is there a reason for this? Same comment for Fig.4 of the reproducibility study vs Fig.1 of the original work. In this latter case the error also seems higher in the reproducibility study (not just the variance). Finally, comparing Table 2 in the reproducibility study and Table 1 in the original work, it seems learning is slower in the reproduced work, especially looking at the accuracy at epoch 50. Can you discuss these differences?

Overall, the reproducibility study does not make specific recommendations to the authors, except to note that it would be a good direction to further explore how to set the hyper-parameter “p” in PADAM. The paper would be improved by a more in-depth discussion of results throughout.

Confidence : 4

@reproducibility-org
Copy link
Collaborator

Hi, please find below a review submitted by one of the reviewers:

Score: 5
Reviewer 2 comment : The authors (of the reproducibility report) attempted to reproduce the cifar-10 and cifar-100 experiments in the Padam paper. The authors plotted training and test curves. However, the training error nor the test error does not seem to match (or even come close the results in the paper). Despite this, the author still concluded that this method is capable of merging the benefit of both Adam and SGD. This conclusion seems to be a little surprising, since there is still a significant gap between the presented training error in the paper and that is reproduced.
Confidence : 4

@reproducibility-org reproducibility-org added the review-complete Review is done by all reviewers label Feb 25, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
checks-complete Submission criteria checks complete review-complete Review is done by all reviewers reviewer-assigned Reviewer has been assigned
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants