-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
test: speed up unit tests by using smaller datasets #563
Conversation
* database.test.cpp: Use a smaller dataset. Speed up locally: 7s -> 800ms * preprocessor.test.cpp: The RSV dataset took 45s -> shift it to the new e2e tests * preprocessor.test.cpp: Use smaller datasets for some tests: 1.8s -> ~500ms
362b1ed
to
84e0d05
Compare
I added GenSpectrum/LAPIS-SILO-e2e#34 to not drop our covered input data |
IMO the dataset is not of immediate importance. In these tests here, it was just some dataset. We already have some other dataset in other tests. |
Exactly. I just wanted to have the issue such that we remember adding it at some point. Not blocking in any way :) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Cool. I haven't looked into all of the details, but looks good to me.
(It's good to have a test suite that covers most of the potential issues when coding, and then have other more expensive tests/test suites that give more certainty for a higher cost. Even in unit tests, could have expensive ones if there are 2 classes of tests, important-but-quick and more-extensive-but-expensive ones.)
We could also learn from other codebases. e.g. duckdb marks suffixes some test-files with .test_slow: |
And then their test runner can execute them based on the arguments given by the developer running the tests |
part of #560
Summary
Local results (debug build):
On CI, time of the unit tests:
PR Checklist
- [ ] All necessary documentation has been adapted or there is an issue to do so.