Polling for Available GPUs when Batch Running SLEAP #777
Replies: 1 comment 4 replies
-
We can't do fractional GPU allocations at the level that SLEAP or TensorFlow operate at since that's a driver-level thing, but one thing that SLEAP does by default is to disable preallocation of GPU memory. What that does is it prevents TensorFlow from using up the entire GPU's memory and instead just grow its usage as needed. This allows you to run other stuff on the GPU at the same time as long as there is sufficient memory available. In terms of auto-selecting a good one to use -- this is a bit trickier and probably a more involved feature enhancement. The easiest way to do this would probably be to parse the output of
A good way to get started might be to do step 1 and add it as a function to |
Beta Was this translation helpful? Give feedback.
-
Hello again SLEAP Devs!
After chatting with a few members of the lab who are trying to run many videos at once, I was wondering how to poll whether or not a GPU is available on the computer and, if it becomes available, start running SLEAP on those instances in parallel if possible. Some of our computers here have 4 or even 5 cards on them and it would be pretty cool to add this sort of functionality to SLEAP if a team/lab doesn't have an orchestrator for spawning jobs available in their cluster. An even cooler enhancement would be to try running fractional GPU jobs from something like this if the card has enough VRAM available, although I don't know if that's easy to configure without tools like Run:AI.
@talmo sent me this section of code for examples that can help make a Python script do it, but I figured I would post about it here and maybe ask about how I could contribute to adding a feature like that to SLEAP one day.
Beta Was this translation helpful? Give feedback.
All reactions