Optimized Cooperative Inference for Energy-Efficient and Low-Latency Mobile Edge Computingopen access
- Authors
- Choi, Hyun-Ho; Lee, Kisong
- Issue Date
- 2025
- Publisher
- IEEE
- Keywords
- confidence thresholds; Cooperative inference; joint optimization; mobile edge computing (MEC)
- Citation
- 2025 International Conference on Information Networking, pp 642 - 647
- Pages
- 6
- Indexed
- FOREIGN
- Journal Title
- 2025 International Conference on Information Networking
- Start Page
- 642
- End Page
- 647
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/58485
- DOI
- 10.1109/ICOIN63865.2025.10993081
- ISSN
- 2996-1572
2996-1580
- Abstract
- To overcome the limitations of standalone inference on edge devices or servers, we propose a cooperative inference method for mobile edge computing (MEC) systems. Using dual confidence thresholds on a small neural network (NN) at the edge, ambiguous images are filtered and sent to a larger NN on the server for reevaluation. We evaluate the method's accuracy, delay, and energy consumption, accounting for confidence score distributions that could trigger false alarms. A joint optimization problem is formulated to minimize delay and energy consumption by selecting optimal confidence thresholds, transmit power, and duty cycle while ensuring accuracy. Experimental results show that this approach significantly reduces delay and energy consumption while achieving higher accuracy than device-only inference and lower costs than server-only inference in various MEC scenarios. © 2025 IEEE.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > Department of Information and Communication Engineering > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.