On energy complexity of fully-connected layers.

Journal: Neural networks : the official journal of the International Neural Network Society
PMID:

Abstract

The massive increase in the size of deep neural networks (DNNs) is accompanied by a significant increase in energy consumption of their hardware implementations which is critical for their widespread deployment in low-power mobile devices. In our previous work, an abstract hardware-independent model of energy complexity for convolutional neural networks (CNNs) has been proposed and experimentally validated. Based on this model, we provide a theoretical analysis of energy complexity related to the computation of a fully-connected layer when its inputs, outputs, and weights are transferred between two kinds of memories (DRAM and Buffer). First, we establish a general lower bound on this energy complexity. Then, we present two dataflows and calculate their energy costs to achieve the corresponding upper bounds. In the case of a partitioned Buffer, we prove by the weak duality theorem from linear programming that the lower and upper bounds coincide up to an additive constant, and therefore establish the optimal energy complexity. Finally, the asymptotically optimal quadratic energy complexity of fully-connected layers is experimentally validated by estimating their energy consumption on the Simba and Eyeriss hardware.

Authors

  • Jiří Šíma
    Institute of Computer Science of the Czech Academy of Sciences, P.O. Box 5, 18207 Prague 8, Czech Republic. Electronic address: sima@cs.cas.cz.
  • Jérémie Cabessa
    Laboratory of Mathematical Economics (LEMMA), Université Paris 2-Panthéon-Assas, 75005 Paris, France.
  • Petra Vidnerová
    The Czech Academy of Sciences, Institute of Computer Science, Pod Vodárenskou věží 271/2, 182 07 Prague 8, Czechia. Electronic address: petra@cs.cas.cz.