You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
About the 'requiring only about three days of training on a single mid-range GPU' mentioned in the paper, there are some doubts I met and hope you can help me resolve:
First of all, I didn't make any changes to the model or the weights and I used the University's Server (A100 GPU) as the step and command mentioned in the ReadMe.
And the problem is that it takes too long time.
Specifically, Action2Motion with two datasets takes many minutes in one epoch, Text2Motion and Unconstrained takes about twenty seconds in one epoch. In other words, I need at least one week or two weeks to train any one of the MDM.
BTW, the loss seems normal, and training records screenshots are attached below. Action2Motion Unconstrained
I also attached my Hardware Configuration screenshots below.
I guess the problem may be the arg.json because I didn't change any resource file.
Looking forward to your early reply~
Thank you!
The text was updated successfully, but these errors were encountered:
Something seems odd here. We tested the code on NVIDIA GeForce RTX 2080 Ti which is significantly weaker on paper compared to your A100, yet, it took about 5GB of memory and runs at about 6.5 iterations/sec.
BTW, I found that the main reason that affects the MDM training speed may be the Sharing Logic of the University Server, I guess?
Just like the screenshots attached show, when I(i.e., PID=2904280) began training the MDM, the `Power Usage` increased by `71W`, and `GPU Memory Usage` increased by `12701MiB`. Does it look normal?
And the speed is still not as quick as expected: for action2motion_humanact12, which is about 20 seconds per epoch; for action2motion_uestc, which is about 3 minutes per epoch; and for unconstrained_humanact12, which is about 20 seconds per epoch. To be honest, it's a bit puzzling for me.
Hi,
About the 'requiring only about three days of training on a single mid-range GPU' mentioned in the paper, there are some doubts I met and hope you can help me resolve:
First of all, I didn't make any changes to the model or the weights and I used the University's Server (A100 GPU) as the step and command mentioned in the ReadMe.
And the problem is that it takes too long time.
Specifically,
Action2Motion
with two datasets takes many minutes in one epoch,Text2Motion
andUnconstrained
takes about twenty seconds in one epoch. In other words, I need at least one week or two weeks to train any one of the MDM.BTW, the loss seems normal, and training records screenshots are attached below.
Action2Motion
Unconstrained
I also attached my Hardware Configuration screenshots below.
I guess the problem may be the arg.json because I didn't change any resource file.
Looking forward to your early reply~
Thank you!
The text was updated successfully, but these errors were encountered: