SUMMARY
UCLA researchers in the Department of Electrical Engineering have developed a novel hardware architecture for computing sparse neural networks.
BACKGROUND
Sparse neural networks (SNNs, or a neural network in which most matrices’ entries are zeros), are used to reduce the number of matrix coefficients, leading to more efficient storage and computations. However, the recently reported NN hardware accelerator cannot compute SNN due to its inability to bypass computation of zero in dataflow. There is a need for neural network architecture that can process and benefit from SNNs.
INNOVATION
Researchers at UCLA have developed an efficient hardware architecture for computing SNNs. The architecture achieves this by using a unique data organization system that allows both zero- and non-zero weights to be stored and later reconstructed using relative address coding. This allows processing of SNNs without the usual downfall of attempting to compute zero in dataflow. This architecture also enables computation of dense networks by a similar process.
POTENTIAL APPLICATIONS
ADVANTAGES
RELATED MATERIALS
PATENT STATUS
United States of America Published Patent Application 20210042610 02/11/2021