Bogacka, K.; Sowinski, P.; Danilenka, A.; Biot, FM.; Wasielewska-Michniewska, K.; Ganzha, M.; Paprzycki, M.... (2024). Flexible Deployment of Machine Learning Inference Pipelines in the Cloud-Edge-IoT Continuum. Electronics. 13(10). https://doi.org/10.3390/electronics13101888
Por favor, use este identificador para citar o enlazar este ítem: http://hdl.handle.net/10251/207596
Título:
|
Flexible Deployment of Machine Learning Inference Pipelines in the Cloud-Edge-IoT Continuum
|
Autor:
|
Bogacka, Karolina
Sowinski, Piotr
Danilenka, Anastasiya
Biot, Francisco Mahedero
Wasielewska-Michniewska, Katarzyna
Ganzha, Maria
Paprzycki, Marcin
Palau Salvador, Carlos Enrique
|
Entidad UPV:
|
Universitat Politècnica de València. Escuela Técnica Superior de Ingenieros de Telecomunicación - Escola Tècnica Superior d'Enginyers de Telecomunicació
|
Fecha difusión:
|
|
Resumen:
|
[EN] Currently, deploying machine learning workloads in the Cloud-Edge-IoT continuum is challenging due to the wide variety of available hardware platforms, stringent performance requirements, and the heterogeneity of the ...[+]
[EN] Currently, deploying machine learning workloads in the Cloud-Edge-IoT continuum is challenging due to the wide variety of available hardware platforms, stringent performance requirements, and the heterogeneity of the workloads themselves. To alleviate this, a novel, flexible approach for machine learning inference is introduced, which is suitable for deployment in diverse environments-including edge devices. The proposed solution has a modular design and is compatible with a wide range of user-defined machine learning pipelines. To improve energy efficiency and scalability, a high-performance communication protocol for inference is propounded, along with a scale-out mechanism based on a load balancer. The inference service plugs into the ASSIST-IoT reference architecture, thus taking advantage of its other components. The solution was evaluated in two scenarios closely emulating real-life use cases, with demanding workloads and requirements constituting several different deployment scenarios. The results from the evaluation show that the proposed software meets the high throughput and low latency of inference requirements of the use cases while effectively adapting to the available hardware. The code and documentation, in addition to the data used in the evaluation, were open-sourced to foster adoption of the solution.
[-]
|
Palabras clave:
|
Machine learning
,
Edge computing
,
IoT
,
Cloud-edge-IoT
,
Inference
,
GRPC
,
Inference server
|
Derechos de uso:
|
Reconocimiento (by)
|
Fuente:
|
Electronics. (eissn:
2079-9292
)
|
DOI:
|
10.3390/electronics13101888
|
Editorial:
|
MDPI AG
|
Versión del editor:
|
https://doi.org/10.3390/electronics13101888
|
Código del Proyecto:
|
info:eu-repo/grantAgreement/EC/H2020/957258/EU/Architecture for Scalable, Self-*, human-centric, Intelligent, Secure, and Tactile next generation IoT/
info:eu-repo/grantAgreement/EC/HE/101069732/EU/Autonomous, scalablE, tRustworthy, intelligent European meta Operating System for the IoT edge-cloud continuum/
|
Agradecimientos:
|
This work was funded by the European Commission, in part under the Horizon 2020
project ASSIST-IoT, grant number 957258. The work of Marcin Paprzycki and Katarzyna WasielewskaMichniewska was funded under the Horizon Europe ...[+]
This work was funded by the European Commission, in part under the Horizon 2020
project ASSIST-IoT, grant number 957258. The work of Marcin Paprzycki and Katarzyna WasielewskaMichniewska was funded under the Horizon Europe project aerOS, grant number 101069732.
[-]
|
Tipo:
|
Artículo
|