Published October 3, 2022 | Version v1
Conference paper Open

Small Reals Representations for Deep Learning at the Edge: A Comparison

  • 1. Università di Pisa
  • 2. MMI spi


The pervasiveness of deep neural networks (DNNs) in edge devices enforces new requirements on information representation. Low precision formats from 16 bits down to 1 or 2 bits have been proposed in the last years. In this paper we aim to illustrate a general view of the possible approaches of optimizing neural networks for DNNs at the edge. In particular we focused on these key points: i) limited non-volatile storage ii) limited volatile memory iii) limited computational power. Furthermore we explored the state-of-the-art of alternative representations for real numbers comparing their performance in recognition and detection tasks, in terms of accuracy and inference time. Finally we present our results using posits in several neural networks and datasets, showing the small accuracy degradation between 32-bit floats and 16-bit (or even 8-bit) posits, comparing the results also against the bfloat family.



Files (312.6 kB)

Name Size Download all
312.6 kB Preview Download

Additional details

Related works

Is identical to
Conference paper: 10.1007/978-3-031-09779-9_8 (DOI)


EPI SGA1 – SGA1 (Specific Grant Agreement 1) OF THE EUROPEAN PROCESSOR INITIATIVE (EPI) 826647
European Commission
TEXTAROSSA – Towards EXtreme scale Technologies and Accelerators for euROhpc hw/Sw Supercomputing Applications for exascale 956831
European Commission