-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Align video recordings with actions #1
Comments
i think this would be solved if we use ttyrec instead of videos. |
How would you apply ttyrec on a regular desktop? Can it handle video games such as Overwatch? |
i did not get the question about "regular desktop" (is it about apps which are not terminal based?) Also at higher frame rates it would be really hard to collect/align the actions, it would be more important to think of a way to handle them from the model-side (eg, think of how CTC loss aligns characters at each timestep). |
No, scaling down the frames is not possible. If you want to try it out, take a screenshot of this page and downscale it by a factor of 2. |
@ClashLuke I'd recommend to read 'Grandmaster level in StarCraft-II using multi agent reinforcement learning'. It has all that you'd need, real time inference with visual input using architecture consisting of tranformers, etc. |
Unfortunately, we can't take one screenshot for every action, as screenshots take 100ms or more. However, recording an entire screen at 60 FPS (the maximum framerate most modern monitors support) is possible. If we later align those frames with the actions taken during a post-processing step, we arrive at roughly the same output without the massive latency overhead. This way, we can retain the capability of allowing the model to "see" what's on the screen.
The text was updated successfully, but these errors were encountered: