Replies: 2 comments 3 replies
-
Hello Mauro,
First of all if you are Linux user you may be lucky in few month. ROCm promise to finally support RNDA(2) till end of 2021. ROCm/ROCm#1180 (comment) This way you'll be able to use it for ML. Also as a stop gap I suggest to check Caffe-OpenCL, also it is as efficient in terms of memory management as Keras/PlaidML - it has quite good performance - also not as fast as cuDNN/miOpen solutions - but it also works (I recommend building it with clBlas support of as BLAS library). Of course if you are on Windows the options are much more limited - as you mentioned PlaidML (that has quite poor performance) . Regarding dlprimitives - it is very young project but what exists seems to be working.
I'd be glad. 1st of all start using it try few things (there are under examples) - try to train some nets. 2nd of course is contributions of code, there are several areas, that each require little bit different knowledge. In general this projects needs one of: knowledge in GPU programming, knowledge of C++ or Python (of course more is better) But it terms of things to do I have a lot:
I agree documentation is lacking, considering it exists only for few month and only recently I proved to myself it actually works well, I think it isn't in that bad shape but need to dig in to get it.
Yes. I want to implement opencl backend for pytorch since this one sees to have much better code base and it is very fast, but it is very complex project as I need to learn lots of PT internals. I did started working to adding DLPrimitives support to caffe-opencl since it is very simple and its code really-really readable (and I'm familiar with it) I do want to extend PlaidML with dlprimitives plaidml/plaidml#1857 but this is probably my last priority unless I'll have git problems with PT. In any case it can be good option as well as it may allow having relatively good performance framework in short time, but I don't see huge community that even willing to answer my question. PlaidML are working to be integrated to TF as backend (since multi-backend-Keras RIP). If they succeed it would be much easier to replace few performance critical ops with dlprimitives at least for channel-first format I support. Regards and remember 1st thing that helps - start using and report issues! (I has written a lot there - probably need to add to blog post) |
Beta Was this translation helpful? Give feedback.
-
Appreciate your effort man, hope other people will notice this repo! |
Beta Was this translation helpful? Give feedback.
-
Hello @artyom-beilis, my name is Mauro, an AI student from Italy.
First thing first, I love the idea behind this project. I am an (un)lucky owner of a 6600xt, a powerful card yet unusable for ML tasks due to the lack of compatibility in the current frameworks. (I tried PlaidML and directML, but c'mon..).
I have thoroughly read your blog post, motivations are strong, but as you might imagine building a community is one of the most important things in the open source world.
What if I wanted to contribute to the project? You have shown all the benchmarks and stuff (and look, results seem great), but
do you plan to include some documentation of the code and some build-related tutorials?
Are you planning to make dlprimitives a tensorflow/pytorch backend?
Thanks for your attention,
Mauro.
Beta Was this translation helpful? Give feedback.
All reactions