TY - GEN
T1 - Time estimation for deep learning model’s inference in distributed processing units
AU - Portugal, Ernesto
AU - Ayala, Angel
AU - Cruz, Francisco
AU - Fernandes, Bruno
AU - Murilo, Sergio
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - One problem with cloud computing is that it may fail to meet the desired time limits for real-time applications. In this regard, fog computing paradigm has gained ground as it complements the cloud by providing nodes with processing and storage capabilities closer to the data generation level. However, this level of the architecture has limited resources, making it necessary to efficiently distribute the workload involved in applications, especially when employing deep learning models. One technique to achieve this is task offloading, which involves distributing inference tasks throughout the architecture. Nevertheless, it is also important to know the time required for these tasks to be carried out within the network in order to obtain the desired response. In this work, we propose a queue-based convolutional neural network that allows estimating the response time for a deep learning inference task. Preliminary results demonstrate a good fit to the behavior of the datasets used in the experiment.
AB - One problem with cloud computing is that it may fail to meet the desired time limits for real-time applications. In this regard, fog computing paradigm has gained ground as it complements the cloud by providing nodes with processing and storage capabilities closer to the data generation level. However, this level of the architecture has limited resources, making it necessary to efficiently distribute the workload involved in applications, especially when employing deep learning models. One technique to achieve this is task offloading, which involves distributing inference tasks throughout the architecture. Nevertheless, it is also important to know the time required for these tasks to be carried out within the network in order to obtain the desired response. In this work, we propose a queue-based convolutional neural network that allows estimating the response time for a deep learning inference task. Preliminary results demonstrate a good fit to the behavior of the datasets used in the experiment.
KW - convolution time series
KW - deep learning
KW - fog computing
KW - time estimation
UR - https://www.scopus.com/pages/publications/85185220356
U2 - 10.1109/LA-CCI58595.2023.10409398
DO - 10.1109/LA-CCI58595.2023.10409398
M3 - Conference contribution
AN - SCOPUS:85185220356
T3 - 2023 IEEE Latin American Conference on Computational Intelligence, LA-CCI 2023
BT - 2023 IEEE Latin American Conference on Computational Intelligence, LA-CCI 2023
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2023 IEEE Latin American Conference on Computational Intelligence, LA-CCI 2023
Y2 - 29 October 2023 through 1 November 2023
ER -