Incentive-Aware Partitioning and Offloading Scheme for Inference Services in Edge Computing
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kim TaeYoung | - |
dc.contributor.author | 김창경 | - |
dc.contributor.author | Lee Seung-seob | - |
dc.contributor.author | Lee Sukyoung | - |
dc.date.accessioned | 2025-03-20T02:50:37Z | - |
dc.date.available | 2025-03-20T02:50:37Z | - |
dc.date.issued | 2024-07 | - |
dc.identifier.issn | 1939-1374 | - |
dc.identifier.uri | https://yscholarhub.yonsei.ac.kr/handle/2021.sw.yonsei/23221 | - |
dc.description.abstract | Owing to remarkable improvements in deep neural networks (DNNs), various computation-intensive and delay-sensitive DNN services have been developed for smart IoT devices. However, employing these services on the devices is challenging due to their limited battery capacity and computational constraints. Although edge computing is proposed as a solution, edge devices cannot meet the performance requirements of DNN services because the majority of IoT applications require simultaneous inference services, and DNN models grow larger. To address this problem, we propose a framework that enables parallel execution of partitioned and offloaded DNN inference services over multiple distributed edge devices. Noteworthy, edge devices are reluctant to process tasks due to their energy consumption. Thus, to provide an incentive mechanism for edge devices, we model the interaction between the edge devices and DNN inference service users as a two-level Stackelberg game. Based on this model, we design the proposed framework to determine the optimal scheduling with a partitioning strategy, aiming to maximize user satisfaction while incentivizing the participation of edge devices. We further derive the Nash equilibrium points in the two levels. The simulation results show that the proposed scheme outperforms other benchmark methods in terms of user satisfaction and profits of edge devices. | - |
dc.format.extent | 13 | - |
dc.publisher | Institute of Electrical and Electronics Engineers | - |
dc.title | Incentive-Aware Partitioning and Offloading Scheme for Inference Services in Edge Computing | - |
dc.type | Article | - |
dc.publisher.location | 미국 | - |
dc.identifier.doi | 10.1109/TSC.2024.3359148 | - |
dc.identifier.wosid | 001290231100025 | - |
dc.identifier.bibliographicCitation | IEEE Transactions on Services Computing, v.17, no.4, pp 1580 - 1592 | - |
dc.citation.title | IEEE Transactions on Services Computing | - |
dc.citation.volume | 17 | - |
dc.citation.number | 4 | - |
dc.citation.startPage | 1580 | - |
dc.citation.endPage | 1592 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
Items in Scholar Hub are protected by copyright, with all rights reserved, unless otherwise indicated.
Yonsei University 50 Yonsei-ro Seodaemun-gu, Seoul, 03722, Republic of Korea1599-1885
© 2021 YONSEI UNIV. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.