Most AI chips and hardware accelerators that power machine learning (ML) and deep learning (DL) applications include floating-point units (FPUs). Algorithms used in neural networks today are often ...
Replacing computationally complex floating-point tensor multiplication with the much simpler integer addition is 20 times more efficient. Together with incoming hardware improvements this promises ...