You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Since Hessian is indefinite for CNNs, CG cannot guarantee the monotonic decrease of errors. So, have you ever tried to use the best step's results instead of the last? In my experiments, the residual errors of r^Tr vary along with the iterations.
The text was updated successfully, but these errors were encountered:
Nope, but I have used simple alternatives like plain gradient descent and steepest descent, which however is also affected by negative alphas, although it performs better than CG in my tests.
I agree that CG does not always work, with some step sizes becoming negative. I am not sure how/if the problem has been addressed in the original paper, but I am open to suggestions and to brainstorm the issue. :)
Since Hessian is indefinite for CNNs, CG cannot guarantee the monotonic decrease of errors. So, have you ever tried to use the best step's results instead of the last? In my experiments, the residual errors of r^Tr vary along with the iterations.
The text was updated successfully, but these errors were encountered: