Skip to content

Latest commit

 

History

History
17 lines (9 loc) · 790 Bytes

File metadata and controls

17 lines (9 loc) · 790 Bytes

Exercise 1: Evaluating Mixed-Precision Performance

I recommend trying out the different mixed-precision choices on your GPU (if you have one):

  • regular float32 ("32-true")

  • regular float16 ("16-true")

  • regular float64 ("64-true")

  • mixed precision with float16 ("16-mixed")

  • mixed-precision with bfloat16 ("bf16-mixed")

For this, you can modify the precision argument of the Trainer in the DistillBert model from lecture 9.1here.

Please also fell free to share your results in the discussion here -- I'd be interested to hear what you find on various GPU architectures!