Tiny Machine Learning has gained significant research interest in the Internet of Things and edge computing fields. In this direction, continuous on-device learning represents an emerging research topic that aims to improve performance and minimize concept drift and privacy violations related to the need to transmit on-field data to a remote server for periodic retraining and refinement of the model. However, the resource consumption resulting from the intensive and prolonged use of the neural network for inference activity, can represent a strong limitation in the use of tiny devices, particularly because of the stringent constraints on memory and power consumption.Therefore, this paper presents the implementation of a novel on-device pruning procedure, designed to reduce the model’s latency and power consumption, without the need to send data to a central server. It runs during the incremental on-board training process, directly on memory-constrained devices based on MCUs without relying on custom hardware. An experimental evaluation on real devices is provided to support machine learning developers in choosing what can be the best compromise between model performance and energy consumption.

On-device training and pruning for energy saving and continuous learning in resource-constrained MCUs

Fusco P.;Rimoli G. P.;Guerriero A.;Palmieri F.;Ficco M.
2026

Abstract

Tiny Machine Learning has gained significant research interest in the Internet of Things and edge computing fields. In this direction, continuous on-device learning represents an emerging research topic that aims to improve performance and minimize concept drift and privacy violations related to the need to transmit on-field data to a remote server for periodic retraining and refinement of the model. However, the resource consumption resulting from the intensive and prolonged use of the neural network for inference activity, can represent a strong limitation in the use of tiny devices, particularly because of the stringent constraints on memory and power consumption.Therefore, this paper presents the implementation of a novel on-device pruning procedure, designed to reduce the model’s latency and power consumption, without the need to send data to a central server. It runs during the incremental on-board training process, directly on memory-constrained devices based on MCUs without relying on custom hardware. An experimental evaluation on real devices is provided to support machine learning developers in choosing what can be the best compromise between model performance and energy consumption.
2026
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11386/4929802
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? 0
social impact