Most Resource Efficient Matrix Vector Multiplication on FPGAs

Lehnert A, Holzinger P, Pfenning S, Müller R, Reichenbach M (2023)


Publication Type: Journal article

Publication year: 2023

Journal

Pages Range: 1-1

DOI: 10.1109/ACCESS.2023.3234622

Abstract

Fast and resource-efficient inference in artificial neural networks (ANNs) is of utmost importance and drives many new developments in the area of new hardware architectures, e.g., by means of systolic arrays or algorithmic optimization such as pruning. In this paper, we present a novel method for lowering the computation effort for ANN inference utilizing ideas from information theory.Weight matrices are sliced into submatrices of logarithmic aspect ratios. These slices are then factorized. This reduces the number of required computations without compromising on fully parallel processing. We create a new hardware architecture for this dedicated purpose. We also provide a tool to map these sliced and factorized matrices efficiently to reconfigurable hardware. By comparing to the state of the art FPGA implementations, we can prove our claim by lowering hardware resources measured in look-up-tables (LUTs) by a factor of three to six. Our method does not rely on any particular property of the weight matrices of the ANN. It works for the general task of multiplying an input vector with a constant matrix and is also suitable for digital signal processing beyond ANNs.

Authors with CRIS profile

Involved external institutions

How to cite

APA:

Lehnert, A., Holzinger, P., Pfenning, S., Müller, R., & Reichenbach, M. (2023). Most Resource Efficient Matrix Vector Multiplication on FPGAs. IEEE Access, 1-1. https://dx.doi.org/10.1109/ACCESS.2023.3234622

MLA:

Lehnert, Alexander, et al. "Most Resource Efficient Matrix Vector Multiplication on FPGAs." IEEE Access (2023): 1-1.

BibTeX: Download