-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add [pinned, unpinnned] matrix to ci_eval.yaml. #767
base: main
Are you sure you want to change the base?
Conversation
|
a0f261c
to
fa7ab97
Compare
If we need more runners, I think we have plenty of extra machines that we can plug in. At this point, with multiple workflows unstable across unpinned versions, my main priority is getting back to a state where we know what works on which versions. When we have more tests that take longer, we can
This matrix approach is a bit inflexible (all variants run whenever the workflow is triggered, individual jobs can't be run in isolation, canceling the job cancels all variants, et c.) but it is simple to understand and add to more workflows. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Not sure about where to run those / the runner resources but overall looks good to me.
Progress on #760. We could make the scheduled jobs test both pinned and unpinned versions like on #767. Cleanup included here: * Dropped the "Installing the PyTorch CPU wheels saves multiple minutes and a lot of bandwidth on runner setup." comments since they are repetitive. Could add them back if people find them useful. * Stopped installing from the root `requirements.txt` in some workflows, instead opting to just install from the more specific `sharktank/requirements-tests.txt` I did not test the changes to scheduled workflows. Could do that on request, or just revert if we see issues.
Progress on #760. Depends on #765.
This changes ci_eval.yml (sharktank perplexity tests for llama) from testing:
to testing:
The pinned versions are shared across the project. We could individually update pins per test workflow or package, but I'd much rather keep the pins shared by default and leave fragmenting as an escape hatch as needed. I can update the other scheduled workflows to match this style if the changes here look good.
Triggered a test run here: https://github.com/nod-ai/shark-ai/actions/runs/12642858744.