You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When trying to apply an average reduction on a tensor filled with float16 elements, we encounter overflow issues. We configure the operation to use float32 as the compute datatype, ensuring that the accumulation occurs in float32, which prevents overflow. Below, I am providing a reproducer of the behavior described.
Hi @kala855, thank you for submitting the issue report. After discussing with our internal teams, we have determined that what you observed is expected behavior. A lower precision data type combined with a large tensor size can result in higher errors and cause overflow issues.
Typically, tensor reduction is part of the tuning process in a machine learning project. It is one of several parameters that can be adjusted to balance precision and performance. Other parameters include tensor size, data types, and more.
If possible, we would appreciate more information about your project context and your specific goals.
When trying to apply an average reduction on a tensor filled with
float16
elements, we encounter overflow issues. We configure the operation to usefloat32
as the compute datatype, ensuring that the accumulation occurs infloat32
, which prevents overflow. Below, I am providing a reproducer of the behavior described.The previous code was tested on mi210 hardware with rocm5.6.1 and using the half library suggested in MIOpen Readme.
Thanks for your help. If there is any workaround or current solution to this, please let me know.
The text was updated successfully, but these errors were encountered: