-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
improve demo/mnist dataProvider speed #713
Conversation
Hi @reyoung |
With GPU enabled, with VGG model, the dataprovider was not bottleneck.
The time of TrainBatch almost be equal to that of forwardBackward, and the that of PyDP2.getNextBatchInternal is hidden by buffer. So use cache with RAM does not work for this model. Following stat are from the second 100Batches
For Cpu enabled, the data provider will not be the bottleneck, since the CPU is slower than GPU for vgg model, so the CPU will be busy for handling forwardbackward. Maybe you should reduce the model complexity to improve some speed if necessary. |
But I also agree with @reyoung , the numpy could be better for handling input data instead of handwritten code. :-) |
The |
@backyes |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, but maybe should be updated before merged.
Because currently Paddle use pre-commit to automatically format python code.
See #874
b0945d1
to
ff4e046
Compare
@reyoung I have used pre-commit to format the code. Please have a look and let me know if there are other problems. Thanks~ |
…addlePaddle#713) * update catalog name to match item name * update catalog name to match item name in windows source compilation
In response to issue #688 , change the dataProvider in MNIST demo to speed up the training.