A cross-border community for researchers with openness, equality and inclusion

ABSTRACT LIBRARY

Analysis of Neural Network Inference Response Times on Embedded Platforms

Publisher: USS

Authors: Huber Patrick, Technical University of Munich;University of Applied Sciences Kempten Göhner Ulrich, University of Applied Sciences Kempten Trapp Mario, Technical University of Munich;Fraunhofer Institute for Cognitive Systems Zender Jonathan, University of Applied Sciences Kempten Lichtenberg Rabea, University of Applied Sciences Kempten

Open Access

  • Favorite
  • Share:

Abstract:

The response time of Artificial Neural Network (ANN)-inference is of utmost importance in embedded applications, particularly continual stream-processing. Predictive maintenance applications require timely predictions of state changes. This study serves to enable the reader to estimate the response time of a given model based on the underlying platform, and emphasises the relevance of benchmarking generic ANN applications on edge devices. We analyse the influence of net parameters, activation functions as well as single- and multi-threading on execution times. Potential side effects such as tact rate variances or other hardware-related influences are being outlined and accounted for. The results underline the complexity of task-partitioning and scheduling strategies while emphasising the necessity of precise concertation of the parameters to achieve optimal performance on any platform. This study shows that cutting-edge frameworks don't necessarily perform the required concertations automatically for all configurations, which may negatively impact performance.

Keywords: Artificial Neural Network Inference,Tensorflow Lite,Embedded Systems,Benchmarking,Response Times

Published in: IEEE Transactions on Antennas and Propagation( Volume: 71, Issue: 4, April 2023)

Page(s): 2908 - 2921

Date of Publication: 2908 - 2921

DOI: 10.1109/TAP.2023.3240032

Publisher: UNITED SOCIETIES OF SCIENCE