Jeong et al., 2018 - Google Patents
IONN: Incremental offloading of neural network computations from mobile devices to edge serversJeong et al., 2018
View PDF- Document ID
- 7854960758113188208
- Author
- Jeong H
- Lee H
- Shin C
- Moon S
- Publication year
- Publication venue
- Proceedings of the ACM symposium on cloud computing
External Links
Snippet
Current wisdom to run computation-intensive deep neural network (DNN) on resource- constrained mobile devices is allowing the mobile clients to make DNN queries to central cloud servers, where the corresponding DNN models are pre-installed. Unfortunately, this …
- 230000001537 neural 0 title abstract description 22
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Programme initiating; Programme switching, e.g. by interrupt
- G06F9/4806—Task transfer initiation or dispatching
- G06F9/4843—Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
- G06F9/4881—Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Programme initiating; Programme switching, e.g. by interrupt
- G06F9/4806—Task transfer initiation or dispatching
- G06F9/4843—Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
- G06F9/485—Task life-cycle, e.g. stopping, restarting, resuming execution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5005—Allocation of resources, e.g. of the central processing unit [CPU] to service a request
- G06F9/5027—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
- G06F9/505—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals considering the load
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5005—Allocation of resources, e.g. of the central processing unit [CPU] to service a request
- G06F9/5011—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5061—Partitioning or combining of resources
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/40—Transformations of program code
- G06F8/41—Compilation
- G06F8/44—Encoding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network-specific arrangements or communication protocols supporting networked applications
- H04L67/10—Network-specific arrangements or communication protocols supporting networked applications in which an application is distributed across nodes in the network
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Jeong et al. | IONN: Incremental offloading of neural network computations from mobile devices to edge servers | |
Bi et al. | Joint optimization of service caching placement and computation offloading in mobile edge computing systems | |
Singh et al. | Accelerating throughput-aware runtime mapping for heterogeneous MPSoCs | |
CN109818786B (en) | Method for optimally selecting distributed multi-resource combined path capable of sensing application of cloud data center | |
CN112291335B (en) | Optimized task scheduling method in mobile edge calculation | |
Wang et al. | SEE: Scheduling early exit for mobile DNN inference during service outage | |
CN104536937A (en) | Big data appliance realizing method based on CPU-GPU heterogeneous cluster | |
US10482210B2 (en) | System, method, and computer readable medium for walking pads: fast power- supply pad-placement optimization | |
Gao et al. | Deep neural network task partitioning and offloading for mobile edge computing | |
Lin et al. | Echo: An edge-centric code offloading system with quality of service guarantee | |
Bohez et al. | Mobile, collaborative augmented reality using cloudlets | |
Shadi et al. | Ready-time partitioning algorithm for computation offloading of workflow applications in mobile cloud computing | |
Zhang et al. | Employ AI to improve AI services: Q-learning based holistic traffic control for distributed co-inference in deep learning | |
Kaya et al. | Seamless computation offloading for mobile applications using an online learning algorithm | |
Mora et al. | Serverless computing at the edge for aiot applications | |
Shi et al. | A framework for mobile cloud computing selective service system | |
US11829799B2 (en) | Distributed resource-aware training of machine learning pipelines | |
JP2019526091A (en) | Method for optimizing an application of a computing system having a plurality of different memory locations interconnected by one or more communication channels, a non-transitory computer readable storage medium including a computer readable program, and a system | |
Bilel et al. | Cunetsim: A gpu based simulation testbed for large scale mobile networks | |
Guan et al. | Design and implementation of offloading and resource management techniques in a mobile cloud environment | |
CN116321199A (en) | Task unloading method, device and medium based on timing diagram and diagram matching theory | |
CN114217881A (en) | Task unloading method and related device | |
AkashKumar | Heuristic for accelerating run-time task mapping in NoC-based heterogeneous MPSoCs | |
Krawczyk et al. | Mobile offloading framework: Solution for optimizing mobile applications using cloud computing | |
Cui | Parallel pso in spark |