Multiple-base Logarithmic Quantization and Application in Reduced Precision AI Computations Cover Image

Multiple-base Logarithmic Quantization and Application in Reduced Precision AI Computations
Multiple-base Logarithmic Quantization and Application in Reduced Precision AI Computations

Author(s): Vassil Dimitrov, Richard Ford, Laurent Imbert, Arjuna Madanayake, Nilan Udayanga, Will Wray
Subject(s): Library and Information Science, Information Architecture
Published by: Институт по математика и информатика - Българска академия на науките

Summary/Abstract: The power of logarithmic quantizations and computations has been recognized as a useful tool in optimizing the performance of large ML models. There are plenty of applications of ML techniques in digital preservation. The accuracy of computations may play a crucial role in the corresponding algorithms. In this article, we provide results that demonstrate significantly better quantization signal-to-noise ratio performance thanks to multiple-base logarithmic number systems (MDLNS) in comparison with the floating point quantizations that use the same number of bits. On a hardware level, we present details about our Xilinx VCU-128 FPGA design for dot product and matrix vector computations. The MDLNS matrix-vector design significantly outperforms equivalent fixed-point binary designs in terms of area (A) and time (T) complexity and power consumption as evidenced by a 4× scaling of AT^2 metric for VLSI performance, and 57% increase in computational throughput per watt compared to fixed-point arithmetic.

  • Issue Year: 2024
  • Issue No: XIV
  • Page Range: 63-70
  • Page Count: 8
  • Language: English
Toggle Accessibility Mode