Decreased precision data types like FP16 increases the performance for compute bound tasks like Matrix Multiplication, Dense Linear Algebra, Convolutions.
May be adding a general abstraction could be usefull for alpaka?
Especially transform and transformReduce functions in alpaka3 seem tobe good candidates for reduced precision types because they're templated and can work with custom data types.
