You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I built and installed the ngraph-bridge with plaidML backend from source and tested execution times. I am not seeing any meaningful speed ups using ngraph with either CPU or PlaidML backend over regular tensorflow (no ngraph).
I am running the classify example suggested in the docs in a loop, so not just measuring the first execution.
ngraph with plaidML: I can see in the activity monitor that GPU utilization spikes up considerably. However, I don't see any speedup compared to no ngraph run.
(iteration # followed by elapsed time for sess.run in seconds)
Hi! I’m from the PlaidML team. We wanted to address the slowness you are seeing when using the PlaidML backend with nGraph.
We’ve been aware of performance issues caused while converting nGraph ops to the PlaidML level. We have actively been working on fixing the most egregious performance issues.
In the longer term, we have also been working on new ways of representing operations in PlaidML which will facilitate even more efficient lowering from nGraph to PlaidML. If you'd like to test out our newer code, you can set the environment variable USE_STRIPE to 1.
Additionally, we'd like to mention a caveat specific to the example you provided:
PlaidML works best with sufficiently large networks and larger batch sizes. This is because GPU-based computations have overhead associated with data transfers between the CPU and GPU that you don't see when you use the CPU by itself. This may be why you don't see a performance benefit out of using your GPU + PlaidML for the inference example you specified.
Hi,
My specs:
iMac 4 Ghz Intel core i7
32 GB DDR3
AMD Radeon R9 M395X 4096 MB
MacOS 10.13.6
I built and installed the ngraph-bridge with plaidML backend from source and tested execution times. I am not seeing any meaningful speed ups using ngraph with either CPU or PlaidML backend over regular tensorflow (no ngraph).
I am running the classify example suggested in the docs in a loop, so not just measuring the first execution.
ngraph with plaidML: I can see in the activity monitor that GPU utilization spikes up considerably. However, I don't see any speedup compared to no ngraph run.
(iteration # followed by elapsed time for sess.run in seconds)
ngraph with CPU backend
No NGraph
Is this the expected result or am I missing something?
The text was updated successfully, but these errors were encountered: