Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Explore solutions for 'portable' performance monitoring with ReFrame #31

Open
casparvl opened this issue Apr 20, 2023 · 0 comments
Open
Assignees

Comments

@casparvl
Copy link
Collaborator

casparvl commented Apr 20, 2023

ReFrame supports performance checking, but this requires hard-coding system name & expected performance in the tests. This breaks test portability. Some ideas would be:

  • Use the standard ReFrame way of comparing against expected performance, but place actual performance references in some file external to the test. These would be system specific, but would at least allow performance checks on known systems, e.g. the ones used in EESSI's own CI.
  • Compare results to known results, with some very large margin. At least this could capture complete mistakes.
    Implement support in ReFrame for outlier detection, instead of hard performance references. This could check e.g. if the current result is within +/- 2 SD of the past 100 results.
  • Implement support in ReFrame for modelling expected performance (e.g. predict expected performance based on core counts, clock speed, processor model...), and then compare to that (again with large margin).
  • Do performance checks externally, based on the ReFrame report.

This task should explore different options, then decide on how to proceed (e.g. maybe we go for option (1) in the short term, but work on other long term options?). Since we planned on co-design with ReFrame, those options are also attractive.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant