←back to thread

85 points homarp | 1 comments | | HN request time: 0.212s | source
Show context
Lerc ◴[] No.44610720[source]
Has there been much research into slightly flawed matrix multiplications?

If you have a measure of correctness, and a measure of performance. Is there a maximum value of correctness per some unit of processing that exists below a full matrix multiply

Obviously it can be done with precision, since that is what floating point is. But is there anything where you can save x% of computation and have fewer than x% incorrect values in a matrix multiplications?

Gradient descent wouldn't really care about a few (Reliably) dud values.

replies(4): >>44610899 #>>44614746 #>>44614820 #>>44617249 #
1. WithinReason ◴[] No.44614746[source]
If you do it in 8-bit it's usually 2x as fast as 16 bit on Tensorcores