1. Introduction
With generations of mobile devices launched by mobile device providers, users focus on implementing more complex application on their MDs. More and more emerging applications (i.e. online game, augmented reality, video optimization/acceleration, real-time monitoring, face recognition) are being integrated to form more powerful mobile application[1,2]. Those powerful mobile applications consist of multiple computation tasks with dependencies. In order to accomplish every computation task, the computation and communication resources of MD will be heavily occupied. However, MD's finite battery life and limited computation capacity pose significant challenges on accommodating the resource demand of those applications[3]. Fortunately, offloading computation tasks partly to Mobile Cloud Computing (MCC)[4] provides a promising technique to the elastic scaling of the capability of MDs. However, driven by the vision of 5G communications, the inherent limitation of MCC, i.e. the long transmission distance from MD to MCC, will lead to exceedingly long latency for mobile applications[5]. Moreover, there still exists many urgent issues to solve, such as increasing demand for high bandwidth, decreasing the energy consumption, and high quality of experience. Mobile Edge Computing (MEC)[6] can address those limitation by offloading computation tasks to the near-user MEC server instead of the remote cloud center.[7] indicated that MEC would play an significant role in 5G, the next generation mobile network. MEC[8,9] is regraded as a brilliant solution to overcome these difficulties.
By combining with partial computation offloading, the computation task can be divided into two parts, which are executed locally and on the MEC server. [10] deeply studied the computation offloading scheme with the two protocols TDMA and OFDMA in a multiuser Mobile Edge Computing Offloading system by joint optimization the offloading ration, computation and communication resources, and formulated the two computation offloading problem as a convex optimization problem and a mixed-integer problem, respectively. Chen et al.[11] studied a multiuser computational offloading problem in a MEC system under the constraints of communication and computation resources, and designed a distributed computation algorithm which could obtain the Nash equilibrium by using the game theoretic. Chen et al.[12] proposed a decentralized computation offloading algorithm which could achieve a Nash equilibrium in a multiple mobile devices MCC system. [13] proposed a computation offloading framework in a multiple wireless access points MEC system by optimizing the chosen edge server, the CPU frequency and computation and communication resources to trade off the energy consumption and execution time of mobile device including multiple independent tasks. [14,15] all carried out reaserch for the purpose of maximizing the computation rate, but the backgrounds of the proposed problem and the strategies and measures show were different. [14] jointly optimized the transmission power at the access point, the task computation mode(local computing or computation offloading) and time allocation to tasks, and derived the optimal solution by using the convex optimization techniques. In order to prolong the standy time of MD, wireless powered transfer(WPT) and MEC had been integrated together. [15] jointly optimized the computation mode selection(local computing or edge computing), the time allocated to wireless power transfer(WPT) and task offloading in a WPT-assisted multi-MD MEC system. And a joint optimization algorithm based on ADMM decomposition technique was proposed to tackle this problem. In [16], the proposed dynamic computation offloading algorithm achieved the multi-objective optimization by jointly optimizing the computation offloading ratio, the CPU frequency allocated to local execution, and the transmission power for computation offloading under the MEC system with an energy harvesting technique. In the following paper[17], MDs not only could offload their jobs to the MEC server, but also could request computation resource from others MD in the same MEC system. Meanwhile, all the MDs obtained energy from the AP by using WPT technique. [17] aimed at minimizing the whole AP's energy transmitted under the constraints of communication and computation resources, the deadline and the power transmission. [18,19] also gived a study on computation offloading. Although these papers are all about independent task with MD as shown previously, there is no dependency relationship among tasks. As the application becomes increasingly complex, multiple tasks, functions or applications are incorporated into an integrated application. Thus, an in-depth study of offloading computation tasks with dependencies is worthwhile.
Coarse-granularity based task offloading policy is used to provide a optimal solution for tasks with dependencies. [20] proposed a collaborative task computation algorithm to prolong the standby time of mobile devices for the application model in the linear topology. [21] provided a comprehensive computation offloading algorithm for computationally intense applications with multiple subtasks to determine which subtasks should be computed in the MD or cloud. [22] aimed to trade-off the energy consumption of MD and latency for application by finding the optimal assignments of tasks executed on local or remote devices. [23] achieved the trade-offs between the energy cost and latency by optimizing the power allocation and the offloading decisions for an application modeled as directed acyclic graph(DAG). [24,25] also could obtain an overall solution by designing a joint scheduling offloading policy for the application with sequential task dependencies. [26] divided the application into a non-offloadable task and multi-offloadable tasks, and a low complexity sub-optimal algorithm was proposed to decide which offloadable task should be transmitted to MEC server. For an application including multiple subtasks depended on each other, [27] proposed a heuristic algorithm based on particle swarm optimizer (PSO) to solve the 0-1 programming problem, where 0 represented local computing and 1 represented edge computing.
Though these studies improve the performance of mobile device by using computation offloading, most of those researches focus on the optimal scheduling order. Meanwhile, the transmission power and the CPU frequency allocated to tasks are fixed. Therefore, we make a study on the computation offloading in the mobile application which includes multiple computational intensive dependent tasks for a single MD MEC system, and propose an algorithm that minimizes the energy cost and improves performance by jointly optimizing the offloading ratio, CPU frequency, the transmission power and the transmission time. [28] compared the performance of offloading the dependent tasks from MD to a remote cloud center or an edge server and made the conclusion that offloading to an edge server is a promising technique in providing better performance than a remote cloud server. [29] took use of offloading strategy which migrated tasks in the task graph to the nearby wearable or mobile device through the available wireless communication interface such as Bluetooth or Wi-Fi. Experimental conclusions of [29] demonstrated that the scheduling policy could achieve the two objectives of extending battery lifetime and enhancing performance. At present, there are not many works related to task graph on MEC system, most of them are sub-task granularity and these works aim at different objective. As far as we know, there is no related work with the goal of minimizing energy consumption of MD by dividing the input data of all the subtasks on task graph bit by bit, which is able to guarantee the higher resource utilization and less energy consumption. Accordingly, this paper aims to minimize the energy consumption of task graph by jointly optimizing the resource allocation strategy.
In this paper, we study an single MD MEC system where MD executes a mobile application made up of multiple computation tasks. The objective of the paper is to minimize the energy consumption of MD with energy-saving resources allocation. In this paper, we model the complex mobile application consisting of multiple interdependent tasks as a DAG task-graph, and develop an energy-efficient computation offloading algorithm by joint optimization the communication and computation resources. Our main contributions are generalized as follows.
Firstly, in term of system model, most the computation offloading algorithms do not take into account the computation results feedback, and partial computation offloading for tasks with dependencies is task granularity rather than dividing the input data bit by bit. Therefore, we propose a system model of partial computation offloading for a application consisting of multiple tasks with dependencies, which divides the computation data by bitwise and integrates results feedback.
Secondly, in term of problem formulation, for tasks with dependencies, most studies are aimed at achieving the objective of saving energy or reducing latency by optimizing scheduling strategy with fixed transmission power and local computation capacity. However, this paper uses the DVFS technique optimize the CPU frequency, and achieves the optimal transmission power. The corresponding time slots for the local computing, uploading the offloading bits and computation results feedback are allocated with a time division mechanism. Therefore, the problem is formulated as an optimization problem that minimizes the MD’s energy consumption with the constraints of the delay, the computation and communication capacity.
Finally, in term of the optimal solution, convex optimization techniques and the Lagrange Multiplier method are used to simplify the problem, and the original non-convex optimization problem is transformed into a univariate nonlinear equation, which accelerates the problem solving. Simulation results demonstrate that the proposed algorithm not only saves energy consumption, but also prolongs the standby time of MD.
The rest of this paper is organized as follows. In Section II, we expounds the system model. The problem is formulated and the algorithm for the problem is proposed in Section III and its performance evaluation and simulation results are analysed and shown in Section IV. The conclusion is given in Section V.
2. System Model
As shown in Fig. 1, we consider a basic two-node MEC system that consists of single MD and one AP node integrated with a MEC server. In order to reduce the impact of multiple wireless channel creation, frequent requests for the MEC server resources, etc. on task offloading, we conduct the study in the single MEC server scenatio. Both of the two nodes are equipped with one single antenna. This system provides the simplex mode. That is to say, the two processes, offloading input data or receiving computation results, can not be executed through the wireless channel at the same time. The MEC server not only provides the same execution environment but has more sufficient resources than the MD. Therefore, the MEC server can accomplish the computation tasks of application more efficient than the MD. The distance is defined as d between MD and AP, and the bandwidth is denoted as B.
Fig. 1. A mobile-edge computing system with a mobile device and an AP
2.1 Computation Application Model
We assume that the MD has an application that is composed of N computation-intensive tasks (See Fig. 1, with an example where N = 6 ). Each task can be offloaded to MEC server by using partial computation offloading technique. This application model can be described by using Directed Acyclic Graph(DAG) G = (V,E) where the node vj in the node set V denotes a task in the application, the edge (vi,vj) represents the dependency from task vi to vj and the edge set E of the DAG denotes the set of dependencies. Seti represents the in-degree set of task i . Task i can not start executing until tasks belonging to Seti are completed.To describe the parametric context of each task, a two-tuple (Ii,ηi) is defined, where i =1,...,N. Accordingly, Ii is the size of task i input data which includes the existing data li and the computation results transmitted from tasks in Seti . According to [10, 17, 19, 20], we learn that there is a linear relationship between the computation result and the input data for the computation task. ηi is the ratio of the output data to the input data for task i . Task i may need the computation results from related tasks if Seti is not null. As such, the total input data size Ii is \(l_{i}+\sum_{k \in S e l_{i}} \eta_{i} l_{i}\) while Seti is non-null. If Seti is an empty set, Ii is equal to li . We use D , a N*N matrix, represent the task graph.
In this paper, we focus on the scenarios where the MD has finite computing capability. In order to accomplish the application, offloading parts of the input data to the MEC server is necessary. We assume that all tasks of the application are executed in a predetermined scheduling policy which results from Breadth-First Traversal(BFT). For example, we can get the scheduling policy (1,2,3,4,5,6 ) in Fig. 1. And then, we realize our objective which minimizes the energy consumption of the MD. Meanwhile, with the given scheduling strategy, we assume that both local computing and computation offloading of the current task begin simultaneously, which ensures that its dependent tasks have completed correctly. λi is the ratio of the offloading data to the input data for task i , so we can get the constraint of λi :
0 ≤ λ i ≤ 1, ∀i ∈ {1,...,N}. (C1)
where λ = {λ1,...,λN} denotes the offloading policy of the tasks graph.
2.2 Communication Model
For each computational intensive task i∈V , its input bits Ii is split into two parts λi Ii ≥ 0 and (1−λi )Ii ≥ 0 bits, which denote the number of bits offloaded to the MEC server through AP and computed locally, respectively. MD receives the offloading computing results from the MEC server. We introduce the TDMA protocol into the system to avoid any interference among tasks. Because of MD and AP equipped with one antenna, the wireless channel can execute only one process during each time slot. As shown in Fig. 2, the protocol divides the whole time T into 2N time slots, and each task contains two time slots ( ).
Fig. 2. The TDMA protocol for mutli-task computation offloading in an application
1) Computation Offloading from MD to the AP
In the time slot, task i offloads λi Ii input bits to the AP with the transmission power ≥0 . We define h2 as the channel power gain between MD and AP. Next, the achievable data rate(bits/sec) for uploading offloading input data from the MD to the AP is defined as
\(r_{i}^{u}=B \log _{2}\left(1+\frac{P_{i}^{t} h^{2}}{N_{0}}\right)\) (1)
where N0 means the additive white Gaussian noise at the receiver of the AP. For simplifying the computation, we define a function \(g(x)=\frac{N_{0}}{h^{2}}\left(2^{\frac{x}{B}}-1\right), x>0\). According to (1), we can get the expression of \(P_{i}^{t}\) with the variable \(r_{i}^{u}\). That is to say, \(P_{i}^{t}=\frac{N_{0}}{h^{2}}\left(2^{\frac{r_{i}^{u}}{B}}-1\right)\). After we get the offloading bits λi Ii and the uploading time \(t_{i}^{u}, r_{i}^{u}\) can be represented as \(r_{i}^{u}=\frac{\lambda_{i} I_{i}}{t_{i}^{u}}\). So \(P_{i}^{t}\) can be re-defined as followed:
\(P_{i}^{t}=\frac{N_{0}}{h^{2}}\left(2^{\frac{I_{i} \lambda_{i}}{B t_{i}^{t}}}-1\right)=g\left(\frac{I_{i} \lambda_{i}}{t_{i}^{u}}\right)\) (2)
Based on the actual observation offered in [30], and the equation recommended by the EARTH project[31], we know that the uploading power \(P_{i}^{t}\) includes the transmission power \(P_{i}^{t}\) and an extra constant circuit power \(P_{t}^{c}\) caused by converting from digital signal to analog signal, packaging, and so on. Moreover, \(P_{i}^{u}\) is linear with \(P_{i}^{t}\). Therefore, we give the following definition:
\(P_{i}^{u}=P_{t}^{c}+\kappa_{t} P_{i}^{t}\) (3)
where κ t denotes the linear growth and is a constant coefficient without unit. According to (3),the energy consumption of uploading data at MD in time slot is expressed as
\(E_{i}^{u}=P_{i}^{u} t_{i}^{u}=\left(g\left(\frac{I_{i} \lambda_{i}}{t_{i}^{u}}\right)+\kappa_{t} P_{i}^{t}\right) t_{i}^{u}\) (4)
2) Downloading the Computation Results from the AP to the MD
Considering that the MD needs to receive the computation results from AP, so the delay cannot be ignored. The computation results of task i are ηi Ii . And then, the achievable data rate for receiving computation results from the AP to the MD is expressed by
\(r_{d}=B \log _{2}\left(1+\frac{P_{F} h^{2}}{N_{0}}\right)\) (5)
where PF denotes the transmission power of the AP, which is a constant. Therefore, rd is a constant. Meanwhile, we have known the amount of the computation results ηi λi Ii . Ultimately, combining (5), we can get the representation as follows:
\(t_{i}^{d}=\frac{\eta_{i} \lambda_{i} I_{i}}{r_{d}}\) (6)
According to [32,33], the MD power consumption \(P_{i}^{d}\) increases linearly with the download data rate rd . So we give the formula of \(P_{i}^{d}\) as follows.
\(P_{i}^{d}=P_{d}^{c}+\kappa_{d} r_{d}\) (7)
where \(P_{d}^{c}\) is similar to the previous definition \(P_{t}^{c}\). However, \(P_{d}^{c}\) is affected by converting from analog signal to digital signal, unpackaging, and so on. κd represents the linearly increasing coefficient. And then, we get the equation of calculating the energy consumed by receiving the computation results from AP in the time slot \(t_{i}^{d}\). By merging (6) and (7), \(E_{i}^{d}\) is expressed as the following equation.
\(E_{i}^{d}=P_{i}^{d} t_{i}^{d}=\left(P_{d}^{c}+\kappa_{d} r_{d}\right) \frac{\eta_{i} \lambda_{i} I_{i}}{r_{d}}=\left(\frac{P_{d}^{c}}{r_{d}}+\kappa_{d}\right) \eta_{i} I_{i} \lambda_{i}\) (8)
2.3 Computing Model
1) Local Computing:
(1− λi)Ii input data of task i is executed on the mobile device(MD) within a duration . C denotes the number of CPU cycles required for computing 1-bit of input data at MD, and fi is the CPU frequency allocated to task i , i.e. \(f_{i}=\frac{\left(1-\lambda_{i}\right) I_{i} C}{t_{i}^{l}}\). In practical application, the CPU frequency of MD has a maximum value, denoted by fmax . So fi is capped by fmax , i.e.
\(\frac{\left(1-\lambda_{i}\right) I_{i} C}{t_{i}^{l}} \leq f_{\max }, \forall i\) (C2)
Also, the computation latency constraint \(t_{i}^{l}\) for local computing of task i must be satisfied as follows:
\(t_{i}^{l} \geq 0\) (9)
Then, the whole local computation tasks must be accomplished before the deadline, i.e.
\(\sum_{i=1}^{N} t_{i}^{l} \leq T\) (C3)
The energy consumption of task i for local computing is
\(E_{i}^{l}=\kappa_{l} f_{i}^{3} t_{i}^{l}=\kappa_{l} \frac{\left(1-\lambda_{i}\right)^{3} I_{i}^{3} C^{3}}{\left(t_{i}^{l}\right)^{2}}, \forall i\) (10)
where κ l means the effective capacitance coefficient which has relation with the chip structure of MD.
According to Fig. 2, for task i , the time that MD starts to transmit input data through wireless channel is simultaneous with starting local execution. Besides, according to (10), the bigger \(t_{i}^{l}\) is, the smaller \(E_{i}^{l}\) is. So we can infer that CPU is not in idle state between the end of the previous task and the beginning of the next task, which means that local execution lasts the whole time period \(t_{i}^{l}\). Consequently, the wireless channel occupation time and local computation time should meet the following time constraint for task i .
\(t_{i}^{u}+t_{i}^{d} \leq t_{i}^{l}, \forall i\) (11)
Replacing \(t_{i}^{d}\) with (6), we can rewrite the former time constraint as below
\(t_{i}^{u}+\frac{\eta_{i} \lambda_{i} I_{i}}{r_{d}} \leq t_{i}^{l}, \forall i\) (C4)
2) Edge-Computation: We assume that the network and computational resources of the MEC server are infinite. Owing to the sufficient computation capability of the MEC server, the delay of receiving offloading data, the time spending on accomplishing the offloaded computation task, and the delay of sending computation results are relatively small and negligible at the MEC server. Therefore,we assume that MD immediately receives computation result from the AP when the offloading data transmission is completed. Since the objective is to minimize the energy consumption of mobile device, the energy consumption of MEC server is negligible.
3. Problem Formulation and Optimal Solution
In this section, we propose a feasible computation offloading scheme to the application modeled as task graph. In order to minimize the energy consumption of MD, the CPU frequency and the transmission power are optimized for local computing and computation offloading, respectively. The solution process of the optimal computation offloading scheme is shown below.
3.1 Problem Formulation
Under the system model above mentioned, we give a trade-off model between the energy consumption of local computing and the energy cost of communication among the AP and the MD to obtain the total minimum energy consumption of the MD. Meanwhile, the trade-off model must satisfy the delay constraint, the limited computation and communication resources. From (4), (8) and (10), we can come to the conclusion that all the energy consumption of MD is impacted by local computing, computation offloading and receiving computation results. Learning from the former section, the local energy consumption is influenced by the locally executed bits (1−λi)Ii and its executed time \(t_{i}^{l}\), and the energy consumed by communication between the AP and the MD is impacted by the time spending on uploading the offloading bits \(t_{i}^{u}\) and the offloading bits λiIi. Then, Let \(\boldsymbol{\lambda}=\left\{\lambda_{1}, \ldots, \lambda_{N}\right\}, \boldsymbol{t}^{u}=\left\{t_{1}^{u}, \ldots, t_{N}^{u}\right\}, \boldsymbol{t}^{l}=\left\{t_{1}^{l}, \ldots, t_{N}^{l}\right\},\) and these three vectors determine how much energy the application will consume. Mathematically, the delay-constrained energy minimization optimization problem is formulated as
\(\begin{aligned} \mathcal{P}: \min _{\left(\boldsymbol{\lambda}, \boldsymbol{t}^{t}, t^{\prime}\right)} & \sum_{i=1}^{N} E_{i}^{u}+E_{i}^{d}+E_{i}^{l} \\ & \text { s.t. } C 1, C 2, C 3, C 4 \\ & C 5: t_{i}^{u}, t_{i}^{l}>0 \quad \forall i \end{aligned}\)
Now we substitute all related equations into problem P, and P is represented as follows.
\(\begin{array}{l} \mathcal{P}_{1}: \min _{\left(\boldsymbol{\lambda}, t^{*}, t^{\prime}\right)} \sum_{i=1}^{N} t_{i}^{u}\left(P_{t}^{c}+\kappa_{t} \frac{N_{0}}{h^{2}}\left(e^{\frac{I_{i} \ln 2 \lambda_{i}}{B r_{i}^{\prime}}}-1\right)\right)+\left(\frac{P_{d}^{c}}{r_{d}}+\kappa_{d}\right) \eta I_{i} \lambda_{i}+\kappa_{l} \frac{\left(\left(1-\lambda_{i}\right) I_{i} C\right)^{3}}{\left(t_{i}^{l}\right)^{2}} \\ \text { s.t. } C 1: 0 \leq \lambda_{i} \leq 1 \quad \forall i \\ C 2: \frac{\left(1-\lambda_{i}\right) I_{i} C}{t_{i}^{l}} \leq f_{\max } \quad \forall i \\ C 3: \sum_{i=1}^{N} t_{i}^{l} \leq T \\ C 4: t_{i}^{u}+\frac{\eta I_{i} \lambda_{i}}{r_{d}} \leq t_{i}^{l} \quad \forall i \\ C 5: t_{i}^{u}, t_{i}^{l}>0 \quad \forall i \end{array}\) (12)
In Problem P1 , C1 denotes the constraint of the offloading ratio, C2 is the maximum CPU frequency constraint for each task of the application, and C3 is the deadline constraint. C4 guarantees that the time cost by local computing must be greater than or equal to the time taken by the data transmission and receiving computation results for task i . Note that P1 is non-convex in general because of the coupling of λi and \(t_{i}^{l}\) in C2. Therefore, we need convert P1 into a convex problem[34] with some measures, and then give the optimal solution.
3.2 Optimal Solution To P1
This subsection gives the optimal solution of P1. To achieve this goal, firstly, we transform C2 into a convex constraint by multiplying both sides of inequality C2 with \(t_{i}^{l}\). The constraint C2 is converted to C2′.
\(\left(1-\lambda_{i}\right) I_{i} C-f_{\max } t_{i}^{l} \leq 0, \forall i.\) (C2′)
Then, P1 is reformulated as follows.
\(\begin{array}{c} \mathcal{P}_{2}: \min _{\left(\boldsymbol{\lambda}, t^{t}, t^{\prime}\right)} \sum_{i=1}^{N} t_{i}^{u}\left(P_{t}^{c}+\kappa_{t} \frac{N_{0}}{h^{2}}\left(e^{\frac{I_{i} \ln 2 \lambda_{i}}{B t_{i}^{n}}}-1\right)\right)+\left(\frac{P_{d}^{c}}{r_{d}}+\kappa_{d}\right) \eta I_{i} \lambda_{i}+\kappa_{l} \frac{\left(\left(1-\lambda_{i}\right) I_{i} C\right)^{3}}{\left(t_{i}^{l}\right)^{2}} \\ \text { s.t. } C 1, C 3, C 4, C 5 \\ \quad C 2^{\prime}:\left(1-\lambda_{j}\right) I_{i} C-f_{\max } t_{i}^{l} \leq 0 \forall i \end{array}\) (13)
Meanwhile, the function \(\frac{\left(1-\lambda_{i}\right)^{3}}{\left(t_{i}^{l}\right)^{2}}\) is convex with respect to 0 < λi < 1 and \(t_{i}^{l}\)>0 [35], and hence the equation \(\frac{\kappa_{l}\left(\left(1-\lambda_{i}\right) I_{i} C\right)^{3}}{\left(t_{i}^{l}\right)^{2}}\) in the objective function is jointly convex with respect to 0 < λi < 1 and 0<\(t_{i}^{l}\)<T. Moreover, the following lemma demonstrates the evidence that Problem P2 is a convex problem.
Lemma 1 Problem P2 is a convex optimization problem.
Proof: Because g(x) is an exponential function and a convex function, its perspective function[13], \(t_{i}^{u} g\left(\frac{I_{i} \lambda_{i}}{t_{i}^{u}}\right)\) is still convex. Also, \(\frac{\kappa_{l}\left(\left(1-\lambda_{i}\right) I_{i} C\right)^{3}}{\left(t_{i}^{l}\right)^{2}}\) is jointly convex function[35]. Thus, the objective function which is the summation of a set of convex functions, holds the convexity. All constraints are the linear convex function. Consequently, we get the result of Lemma 1.
Assume that Problem P2 is feasible, we use the Lagrange Multiplier method to simply the problem solving. We define αi≥0, βi≥0, γi≥0 as the Lagrangian multipliers for C1, C2′ ,C4, respectively. µ is the Lagrange multiplier for C3. Therefore, the partial Lagrangian function of P2 is given by
\(\begin{aligned} \mathcal{L}\left(\boldsymbol{\lambda}, \boldsymbol{t}^{u}, \boldsymbol{t}^{l}, \boldsymbol{\alpha}, \boldsymbol{\beta}, \boldsymbol{\gamma}, \mu\right) &=\sum_{i=1}^{N} t_{i}^{u}\left(P_{t}^{c}+\frac{\kappa_{t} N_{0}}{h^{2}}\left(e^{\frac{I_{i} \ln 2 \lambda_{i}}{B t_{i}^{n}}}-1\right)\right)+\left(\frac{P_{d}^{c}}{r_{d}}+\kappa_{d}\right) \eta I_{i} \lambda_{i}+\kappa_{l} \frac{\left(\left(1-\lambda_{i}\right) I_{i} C\right)^{3}}{\left(t_{i}^{l}\right)^{2}} \\ &+\sum_{i=1}^{N} \alpha_{i}\left(\lambda_{i}-1\right)+\sum_{i=1}^{N} \beta_{i}\left(\left(1-\lambda_{i}\right) I_{i} C-f_{\max } t_{i}^{l}\right)+\sum_{i=1}^{N} \gamma_{i}\left(t_{i}^{u}+\frac{\eta_{i} I_{i} \lambda_{i}}{r_{d}}-t_{i}^{l}\right) \\ &+\mu\left(\sum_{i=1}^{N} t_{i}^{l}-T\right) \end{aligned}\) (14)
Let {\(\lambda_{i}^{*}, t_{i}^{u^{*}}, t_{i}^{l^{*}}\)} denote the optimal solution of P2 . The optimal solution always satisfies all constraints and has the minimum energy consumption of MD. By applying KKT conditions, the following necessary and sufficient conditions must be true.
\(\frac{\partial \mathcal{L}}{\partial \lambda_{i}^{*}}=\frac{\kappa_{t} N_{0} I_{i} \ln 2}{B h^{2}} e^{\frac{I_{i} \ln 2 \lambda_{i}^{*}}{B t_{i}^{n^{*}}}}+\frac{\left(P_{d}^{c}+\kappa_{d} r_{d}\right) \eta_{i} I_{i}}{r_{d}}-\frac{3 \kappa_{l}\left(1-\lambda_{i}^{*}\right)^{2} I_{i}^{3} C^{3}}{\left(t_{i}^{l^{*}}\right)^{2}}+\alpha_{i}^{*}-\beta_{i}^{*} I_{i} C+\gamma_{i}^{*} \frac{\eta_{i} I_{i}}{r_{d}}=0\) (a)
\(\frac{\partial \mathcal{L}}{\partial t_{i}^{u^{*}}}=P_{t}^{c}+\kappa_{t} \frac{N_{0}}{h^{2}}\left(e^{\frac{I_{i} \ln 2 x_{i}^{*}}{B i_{i}^{*}}}-1\right)-\kappa_{t} \frac{n_{0}}{h^{2}} \frac{I_{i} \ln 2 \cdot \lambda_{i}^{*}}{B \cdot t_{i}^{u^{*}}} e^{\frac{I_{i} \ln 2 \cdot \lambda_{i}^{*}}{B \cdot t^{\prime}}}+\gamma_{i}^{*}=0\) (b)
\(\frac{\partial \mathcal{L}}{\partial t_{i}^{l^{*}}}=-\frac{2 \kappa_{l}\left(1-\lambda_{i}^{*}\right)^{3} I_{i}^{3} C^{3}}{\left(t_{i}^{t^{*}}\right)^{3}}-\beta_{i}^{*} f_{\max }-\gamma_{i}^{*}+\mu^{*}=0\) (c)
\(\alpha_{i}^{*}\left(\lambda_{i}^{*}-1\right)=0, \forall i\) (d)
\(\beta_{i}^{*}\left(\left(1-\lambda_{i}^{*}\right) I_{i} C-f_{\max } t_{i}^{l^{*}}\right)=0, \forall i\) (e)
\(\gamma_{i}^{*}\left(t_{i}^{u^{*}}+\frac{\eta_{i} I_{i} \lambda_{i}}{r_{d}}-t_{i}^{l *}\right)=0, \forall i\) (f)
\(\mu^{*}\left(\sum_{i=1}^{N} t_{i}^{l^{*}}-T\right)=0\) (g)
According to (b), we can get Lemma 2 which describes the relations among {\(\lambda_{i}^{*}, u_{i}^{u^{*}}, \gamma_{i}^{*}\)}.
Lemma 2 The optimal {\(\lambda_{i}^{*}, u_{i}^{u^{*}}, \gamma_{i}^{*}\)} satisfies
\(\frac{\lambda_{i}^{*}}{t_{i}^{u^{*}}}=\frac{B}{I_{i} \ln 2}\left(1+W\left(\frac{-1}{e}+\frac{\left(P_{t}^{c}+\gamma_{i}\right) h^{2}}{\kappa_{t} e N_{0}}\right)\right) \forall i,\) (15)
where W(x) defines the Lambert-W function, i.e. x=W(x)ew(x).
Proof: With the help of (b), we have \(\left(\frac{I_{i} \ln 2 \cdot \lambda_{i}^{*}}{B \cdot t_{i}^{u^{*}}}-1\right) e^{\left.\frac{I_{i} \ln 2 \cdot \lambda_{i}^{*}}{B \cdot t_{i}^{*^{*}}}-1\right)}=\frac{-1}{e}+\frac{\left(P_{t}^{c}+\gamma_{i}^{*}\right) h^{2}}{\kappa_{t} e N_{0}}\). According to the Lambert-W function, i.e. x=W(x)ew(x), which is the inverse function of f(x) =xex. We can straightforwardly infer that \(\left(\frac{I_{i} \ln 2 \cdot \lambda_{i}^{*}}{B \cdot t_{i}^{u^{*}}}-1\right)=W\left(\frac{-1}{e}+\frac{\left(P_{t}^{c}+\gamma_{i}^{*}\right) h^{2}}{\kappa_{t} e N_{0}}\right),\) which leads to the result in Lemma 2 with some simple operations.
With some operations on (c), we can get the following expression. \(\mu^{*}=\frac{2 \kappa_{l}\left(1-\lambda_{i}^{*}\right)^{3} I_{i}^{3} C^{3}}{\left(t_{i}^{t^{*}}\right)^{3}}+\beta_{i}^{*} f_{\max }+\gamma_{i}^{*}, \forall i\). Therefore, we can get the result µ* > 0 . Combining with (g), the summation of \(t_{i}^{l^{*}}\) is identically equal to T , i.e.
\(\sum_{i=1}^{N} t_{i}^{l^{*}}=T\) (16)
which conforms to the monotonically decreasing property of the objective function respect to \(t_{i}^{l}\). According to (a), (c), (d), we can get the following Lemma 3. In order to reduce the complexity of interpretation, we firstly define
\(\varphi_{i}\left(\gamma_{i}\right)=1+W\left(\frac{-1}{e}+\frac{\left(P_{t}^{c}+\gamma_{i}\right) h^{2}}{\kappa_{t} e N_{0}}\right).\) (17)
Lemma 3 The optimal {\(\lambda_{i}^{*}, t_{i}^{*}, \gamma_{i}^{*}\)} satisfies
\(\frac{1-\lambda_{i}^{*}}{t_{i}^{t^{*}}}=\left(\frac{\frac{\kappa_{t} N_{0} \ln 2}{B h^{2}} e^{\varphi_{i}\left(y_{i}^{*}\right)}+\frac{\eta_{i} \gamma_{i}^{*}}{r_{d}}+\left(P_{d}^{c}+\kappa_{d} r_{d}\right) \frac{\eta_{i}}{r_{d}}}{3 \kappa_{l} I_{i}^{2} C^{3}}\right)^{\frac{1}{2}}\) (18)
The ratio is influenced by the i-th task parameters of {\(\eta_{i}, I_{i}\)} and the lagrangian multiplier γi.
Proof: At first, according to (d), because not all the input data is transmitted to the MEC server, we obtain \(\alpha_{i}^{*}\). The result makes sure that the local energy cost is not the maximum. And then, to ensure (e) true, since the CPU frequency fi allocated to task i always less than fmax, we obtain \(\beta_{i}^{*}\)=0. Plugging \(\beta_{i}^{*}\)=0 into (c), and making some simple manipulations, we obtain the following result.
\(\mu^{*}=\frac{2 \kappa_{l}\left(1-\lambda_{i}^{*}\right)^{3} I_{i}^{3} C^{3}}{\left(t_{i}^{*}\right)^{3}}+\gamma_{i}^{*}, \forall i . \) (19)
At last, we substitute (17), \(\alpha_{i}^{*}\)=0, \(\beta_{i}^{*}\)=0 into (a). The Lemma 3 is inferred.
We know \(f_{i}=\frac{\left(1-\lambda_{i}\right) I_{i} C}{t_{i}^{l}} \leq f_{\max }\). Meanwhile, plugging (18) into fi, we obtain an inequality about \(\gamma_{i}^{*}\). We introduce the function \(F_{i}\left(\gamma_{i}^{*}\right)\) for replacing the ratio \(\frac{1-\lambda_{i}^{*}}{t_{i}^{t^{*}}}\), i.e. Correspondingly, the range of \(\gamma_{i}^{*}\) can be described as Proposition 1.
Proposition 1 \(F_{i}\left(\gamma_{i}^{*}\right)\) is a monotonically increasing function when \(\gamma_{i}^{*}\)>0. The one and only \(\gamma_{i}\)>0 satisfies \(F_{i}\left(\gamma_{i}^{*}\right)=f_{\max }\) existed. The value is defined as \(\gamma_{i}^{(0)}\), and \(\gamma_{i}^{*} \leq \gamma_{i}^{(0)}\).
Proof: eϕi(γi) is an increasing function with γi ≥ 0 because ϕi (γi) is an increasing function with \(\gamma_{i} \geq 0\). Meanwhile, \(\frac{\eta_{i}}{r_{d}} \gamma_{i}\) is a linear function. Because Fi(γi) is the square root of the sum of an increasing function, we infer that Fi(γi) is an increasing function with γi ≥ 0 . γi → +∞ and Fi(γi) → +∞. So we can get the unique by using the Bi-Section Search method. The Proposition 1 is proved.
\(\mu^{*}=\gamma_{i}^{*}+2 \kappa_{l}\left(F_{i}\left(\gamma_{i}^{*}\right)\right)^{3}.\) (20)
Now, substituting (18) into (19), we get (20). Learning from (18), we acknowledge that µ is associated with \(\gamma_{i}^{*}\), ηi, and the performance parameters of SD. Although the ratio of (18) changes over ηi, Ii, \(\gamma_{i}^{*}\), the lagrangian multiplier µ is fixed. Meanwhile, µ is a scalar rather than a vector. Therefore, the lagrangian multiplier \(\gamma_{i}^{*}\) should satisfy \(\gamma_{i}^{*}\)>0. In order to meet the condition (f), we get the result \(t_{i}^{u^{*}}+\frac{\eta_{i} I_{i} \lambda_{i}^{*}}{r_{d}}=t_{i}^{t^{*}}\) because of \(\gamma_{i}^{*}\)>0. Accordingly, we obtain Lemma 4 by using the above derivation results, which denotes that the local execution time \(t_{i}^{l^{*}}\) can be expressed as a function with a single variable lagrangian multiplier * µ .
Lemma 4 The optimal \(t_{i}^{l^{*}}\) can be expressed as
\(t_{i}^{l^{*}}=Q_{i}\left(\mu^{*}\right)=\frac{1}{\frac{r_{d}}{I_{i} \eta_{i}}-\frac{r_{d}^{2} \ln 2}{\eta_{i}\left(I_{i} r_{d} \ln 2+B I_{i} \eta_{i} \varphi_{i}\left(\psi_{i}^{-1}\left(\mu^{*}\right)\right)\right)}+\frac{\left(\mu^{*}-\psi_{i}^{-1}\left(\mu^{*}\right)\right)^{\frac{1}{3}}}{I_{i} C\left(2 \kappa_{l}\right)^{\frac{1}{3}}}},\) (21)
Qi(µ) is monotonically decreasing with µl≤ µ*≤ µr. If , and tu do not exist. \(\sum Q_{i}\left(\mu^{l}\right)<T, \lambda,t^{\prime}\), and tu do not exist. If \(T<\sum Q_{i}\left(\mu^{l}\right)\) and \(\sum Q_{i}\left(\mu^{r}\right)<T, \lambda, \boldsymbol{t}^{l}\), and tu are unique.
Proof: Firstly, \(\mu=\psi_{i}\left(\gamma_{i}\right)\) is introduced to simplify the derivation and computation, i.e. \(\psi_{i}\left(\gamma_{i}\right)=\gamma_{i}+2 \kappa_{l} F_{i}^{3}\left(\gamma_{i}\right) \cdot \psi_{i}^{\prime}\left(\gamma_{i}\right)=1+6 \kappa_{l} F_{i}^{2}\left(\gamma_{i}\right) F_{i}^{\prime}\left(\gamma_{i}\right)\) denotes its derivation. Learning from Proposition 1, we can get \(\psi_{i}^{\prime}\left(\gamma_{i}\right)>1\) because Fi (γi) is a monotonically increasing function \(F_{i}^{\prime}\left(\gamma_{i}\right)>0\). Consequently, ψi(γi) is a monotonically increasing function with γi >0 . Given
the value of µ , we can get a unique γi , and vice versa. The inverse function of ψi(γi) is denoted as \(\gamma_{i}=\psi_{i}^{-1}(\mu)\), and Fi(γi) can be rewritten as \(F_{i}\left(\gamma_{i}\right)=\left(\frac{\psi_{i}\left(\gamma_{i}\right)-\gamma_{i}}{2 \kappa_{l}}\right)^{\frac{1}{3}}\). As we all know, the derivative of the original function is equal to the inverse of the derivative of its inverse function in calculus. Hence, \(0<\left(\psi_{i}^{-1}\right)^{\prime}(\mu)=\frac{1}{\psi_{i}^{\prime}\left(\gamma_{i}\right)}<1\), where µ= ψi(γi). Learning from Lemma 2 and Lemma 3, we can get the expression of \(t_{i}^{l^{*}}\) by solving the simultaneous equation (22). We use \(\psi_{i}^{-1}(\mu)\) replace the Lagrangian multiplier γi in (23), and we can use a function Qi(µ*) only associated with µ* represent \(t_{i}^{l^{*}}\), which is (21).
\(\left\{\begin{array}{l} t_{i}^{u^{*}}+\frac{\eta_{i} I_{i} \lambda_{i}^{*}}{r_{d}}=t_{i}^{*} \\ (15),(18) \end{array}\right.\) (22)
\(t_{i}^{l^{*}}=\frac{1}{\frac{r_{d}}{I_{i} \eta_{i}}-\frac{r_{d}^{2} \ln 2}{I_{i} \eta_{i}\left(r_{d} \ln 2+B \eta_{i} \varphi_{i}\left(\gamma_{i}^{*}\right)\right)}+\frac{F_{i}\left(\gamma_{i}^{*}\right)}{I_{i} C}}\) (23)
Secondly, based on the property of Lambert-W function \(\varphi_{i}^{\prime}>0\), and \(0<\left(\psi_{i}^{-1}\right)^{\prime}<1\), the derived function \(Q_{i}^{\prime}(\mu)<0\), and Qi(µ) is µ 's rigid monotony decrease by degrees function.
\(Q_{i}^{\prime}(\mu)=\frac{\frac{B I_{i} r_{d}^{2} \ln 2 \varphi_{i}^{\prime}\left(\psi_{i}^{-1}(\mu)\right)\left(\psi_{i}^{-1}(\mu)\right)^{\prime}}{\left(I_{i} r \ln 2+B I_{i} \eta_{i} \varphi\left(\psi_{i}^{-1}(\mu)\right)\right)^{2}}+\frac{1}{3 I_{i} C\left(2 \kappa_{l}\right)^{\frac{1}{3}}} \cdot \frac{1-\left(\psi_{i}^{-1}(\mu)\right)^{\prime}}{\left(\mu-\psi_{i}^{-1}(\mu)\right)^{\frac{2}{3}}}}{-\left(\frac{r}{I_{i} \eta_{i}}-\frac{r_{d}^{2} \ln 2}{\eta_{i}\left(I_{i} r_{d} \ln 2+B I_{i} \eta_{i} \varphi\left(\psi_{i}^{-1}(\mu)\right)\right)}+\frac{\left(\mu-\psi_{i}^{-1}(\mu)\right)^{\frac{1}{3}}}{I_{i} C\left(2 \kappa_{l}\right)^{\frac{1}{3}}}\right)^{2}}<0\) (24)
Thirdly, we get the interval range of \(\gamma_{i}^{*} \text { as } 0 \leq \gamma_{i}^{*} \leq \gamma_{i}^{(0)}\) according to Proposition 1. And, (20) indicates that µ is a monotonically increasing function with γi . We define \(\mu_{l}^{(0)}=\max \left\{\psi_{i}(0)\right\}\) and \(\mu_{r}^{(0)}=\min \left\{\psi_{i}\left(\gamma_{i}^{(0)}\right)\right\}\), and the numeric zone of µ* is displayed as follows.
\(\mu_{l}^{(0)} \leq \mu^{*} \leq \mu_{r}^{(0)}.\) (25)
Meanwhile, according to (21), is monotonically decreasing with µ* ≥ 0 . What’s more, \(0 \leq t_{i}^{*} \leq T\). By solving \(t_{i}^{l^{*}}=T\), we can get a unique value of µ , defined as \(\mu_{l}^{(1)}\). Let \(\mu^{l}=\max \left(\mu_{l}^{(0)}, \mu_{l}^{(1)}\right)\) and \(\mu^{r}=\mu_{r}^{(0)}\). The restriction of µ* is rewritten as
\(\mu^{l} \leq \mu^{*} \leq \mu^{r}.\) (26)
By using the well studied methods, i.e. Bi-Section Search, Newton Method, solve µi =ψi(γi) = ∀i , the numeric zone of γi, ∀i can be achieved with the given lower and upper limits of µ .
Ultimately, by judging the value of \(\sum Q_{i}(\mu)\), i.e. \(\sum t_{i}^{l}\), we can determine whether problem P2 exists the optimal solution λ*, tl*, and tu* or not. If \(\sum Q_{i}\left(\mu^{r}\right)>T \text { or } \sum Q_{i}\left(\mu^{l}\right)<T\), it does not exist the optimal solution. If \(T<\sum Q_{i}\left(\mu^{l}\right)\) and \(\sum Q_{i}\left(\mu^{r}\right)<T\), the unique optimal solution can be achieved by solving (16). Lemma 4 is proved now.
Accordingly, problem P2 , a multiple variables non-linear optimization problem with multi-constraint, is equivalently transferred to a non-linear equation with a linear inequality constraint. The nonlinear equation is expressed as follows:
\(\begin{array}{l} \mathcal{P}_{3}: \sum_{i=1}^{N} Q_{i}(\mu)=T \\ \text { s.t. } \mu^{l} \leq \mu \leq \mu^{r} \end{array}\) (27)
Since \(t_{i}^{l}\) is a monotonic decreasing function respect to µ , \(\sum t_{i}^{l}\) inherits the descending characteristic. Therefore, there exists a unique µ to P3 . Then, we propose our algorithm Bi-JOTD to calculate the optimal solution to P3 . The solution is the optimal offloading strategy for P1 , too. The proposed optimization method not only successfully avoids solving a nonlinear convex optimization problem with high dimensions and high orders, but also overwhelmingly decreases on the computational complexity and the computation time to acquire the optimal strategy. The following pseudo-code gives the accurate processes to solve the non-linear equation with a linear inequality constraint.
Algorithm 1 calculates the approximate multiplier γi which satisfies (19) with given µ . Bi-JOTD computes the multiplier µ among the linear constraint. After Lagrangian multipliers and µ* are achieved, we can use (15), (18), (21) obtain the final optimal resource allocation strategy λ*, tl*, and tu*.
Through analyzing the proposed algorithm Bi-JOTD, the complexity of Bi-JOTD is associated with the iteration number of inner and outer Bi-Section Search algorithms. We define Cinner and Couter as the complexity of inner and outer algorithms, respectively. Therefore, the complexity of Bi-JOTD algorithm is O(NCinner Couter) . Compared with the original Problem P, our proposed algorithm can greatly reduce the complexity.
4. Simulation Results
In this section, the simulation results of the proposed computation offloading scheme through joint optimization time allocation and dependency tasks is presented. The scheme is named as Bi-JOTD in the simulation results. Based on [16, 23, 24], the simulation parameters are set as follows. We assume that the channel reciprocity holds for the downlink and uplink, and thus \(h_{u}^{2}=h_{d}^{2}=h^{2}\). The channel power gain is modeled as \(h^{2}=10^{-3} d^{-\zeta} \Phi\)[24], where Φ stands for the short-term fading which obeys the Rayleigh fading. For distance d in meters with the same path-loss exponent ζ , a 30dB average signal power attenuation is assumed for all the channels at reference of 1m. Moreover, the input data size and the ratio of the computation results follow the uniform distribution with Ii ∈ [10,100]KB ∈ and ηi ∈ [0.01,0.1] , separately. The other detailed parameters are listed in Table 1.
Table 1. Simulation Parameters
Meanwhile, in order to comparison, the following five baselines[5,16,17,23,24,25,26], are displayed.
1) λ = 0.4 : The computation input data are divided into two parts for each task. The ratio 0.4 λi = of input data is accomplished in the MEC server, and the rest of input data is completed in MD.
2) Task: In the task-graph representing the mobile application, some tasks are offloaded to MEC server, and the rest of tasks execute on the mobile device.
3) λ = 0.8 : The offloading ratio of computation data transmitted to MEC server is set as λi = 0.8.
4) Ave-JOTD: The offloading ratio λi is set as the average of suboptimal solution \(\lambda_{i}\), i.e. \(\lambda_{i}=\overline{\sum \lambda_{i}^{*}}\).
5) PSO-JOTD: The suboptimal task offloading scheme is achieved by using particle swarm optimizer (PSO).
4.1 The Validation of the Optimal Solution
In this subsection, we verify the correctness and superiority of the proposed algorithm. The energy consumption of the approximate solution is close to the cost of the exact solution in a tolerable accuracy range, which compares the simulation results of different block input data I KB ( ) and the ratio η at the same deadline T s =1.0 in Fig. 3 and Fig. 4, respectively.
Fig. 3. The energy consumption with different I versus T
Fig. 4. The energy consumption with different η versus T
By comparing with the energy consumption between the global search with the default step size of 0.001 among the fixed range of offloading ratio λ and the proposed optimization algorithm, the energy consumption of the proposed approximate solution is minimum. The simulation results of Fig. 3 and Fig. 4 provide the evidence that problem P2 is a convex optimization problem. Meanwhile, the process of solving P2 is correct
4.2 The Simulation Results for Task Graph
This subsection evaluates the energy consumption of the proposed Bi-JOTD algorithm and other five baselines, which spends on computing and communicating versus the deadline T Fig. 5 makes the comparison simulation of energy consumption with a vector (Ii, ηi) , where Ii is fixed and ηi is random, and Fig. 6 gives a simulation with random Ii and fixed ηi. The proposed Bi-JOTD algorithm is obviously superior to the other baselines. Specifically, the energy consumption of the proposed algorithm is always less than other baselines, which further displays the advantages of offloading the computation tasks to the MEC server. The minimum energy cost has decreased trend with the increase of T . Therefore, Bi-JOTD algorithm has some advantages in solving multi-task application. Then, making a comparison test between Bi-JOTD and Ave-JOTD, we can see that the energy cost of Bi-JOTD is less than Ave-JOTD with the increase of T . It can be concluded that task's result feedback has a great influence on the ratio of offloading input data. Hence, we should comprehensively take task's feedback into account in the task graph to solve the objective of the formulated problem..
Fig. 5. The energy consumption with fixed I and random η versus T
Fig. 6. The energy consumption with fixed η and random I versus T
This subsection evaluates the energy consumption of the proposed Bi-JOTD algorithm and other five baselines, which spends on computing and communicating versus the deadline T .
The figure simulated in Fig. 7 is based on 300 times simulated testing, in which i I , ηi are randomly selected on the basis of the above assumptions in every implementation, modeling the actual computation offloading scenario. We simulate the resource allocation, and the statistics of success ratios are brought in Fig. 7. The success ratios of all the schemes increase with T , as we excepted. Besides, the success ratio of Bi-JOTD grows faster than others. Bi-JOTD can achieve a high success ratio even with low latency T . Meanwhile, as time goes by, the success ratio of Bi-JOTD reaches 100% and remains stable. Compared with other baselines, our proposed algorithm has more advantages in a high success ratio. Fig. 8 shows the relationship of the minimum energy consumption of MD, the input data and the ratio of computation results. The energy consumption is achieved by using our proposed algorithm, and T s =1.5 . According to Fig. 8, while setting the value of η , the energy consumption increases significantly as the input data increases. When setting the value of the input data I , the energy consumption also shows an increasing trend with the increase of η , but the growth is not very obvious. In particular, the application cannot be accomplished while I KB =100 and η=0.1 . Integrated with the previous simulation results and the optimization processes, the proposed algorithm not only has a great advantage in energy consumption, but also has a good performance on the executability of the computation offloading strategy. Ultimately, combining all figures and the equivalence of problem P1 , P2 , and P3 , we can get the conclusion that our proposed algorithm not only can obtain higher efficiency, less energy consumption and better performance, but also can be calculated more easily than others.
Fig. 7. The succeeding assignment ratios with random η and I versus T
Fig. 8. shows the relationship of the minimum energy consumption of MD, the input data I and the ratio of the computation results η
5. Conclusion
This work studies a computation offloading scheme for an application including multiple computation components with dependencies under a single-user MEC system. MEC server assists MD in completing its computation-intensive latency-critical application. The application is modeled as a DAG task-graph. By jointly optimizing the offloading ratio, CPU frequency, transmission power and transmission time, the objective that minimizes the energy consumption of MD is achieved. The problem is formulated as a optimization problem. The nonlinear equation with a linear inequality is obtained by using the Lagrange Multiplier method and convex optimization techniques. A double Bi-Section Search algorithm is proposed to solve the transformed nonlinear equation. Simulation results reveal that the proposed strategy greatly reduces the energy consumption while compared with the used baselines. The proposed computation offloading scheme not only reduces the difficulty of problem solving, but also prolongs the MD’s standy time and achieves better performance. Meanwhile, we will extend our study of using partial computation offloading strategy to deal with the application including dependent tasks to the multiple mobile devices multiple MEC servers scenario.
References
- Z. Q. Jaber and M. I. Younis, "Design and implementation of real time face recognition system (rtfrs)," International Journal of Computer Applications, vol. 94, no. 12, pp. 15-22, 2014. https://doi.org/10.5120/16395-6014
- J. Kephart and D. Chess, "The vision of autonomic computing," Computer, vol. 36, no. 1, pp. 41-50, Jan. 2003. https://doi.org/10.1109/MC.2003.1160055
- K. Kumar and Y. H. Lu, "Cloud computing for mobile users: Can offloading computation save energy?," Computer, vol. 43, pp. 51-56, Apr. 2010. https://doi.org/10.1109/MC.2010.98
- N. Vallina-Rodriguez and J. Crowcroft, "Energy management techniques in modern mobile handsets," IEEE Communications Surveys & Tutorials, vol. 15, no. 1, pp. 179-198, First 2013. https://doi.org/10.1109/SURV.2012.021312.00045
- Y. Mao, C. You, J. Zhang, K. Huang, and K. B. Letaief, "A survey on mobile edge computing: The communication perspective," IEEE Communications Surveys & Tutorials, vol. 19, no. 4, pp. 2322-2358, 2017. https://doi.org/10.1109/COMST.2017.2745201
- M. Patel, B. Naughton, C. Chan, N. Sprecher, S. Abeta, A. Neal et al., "Mobile-edge computing introductory technical white paper," ETSI, Sophia Antipolis, France, and MEC, London, U.K., Tech. Rep., pp. 1089-7801, 2014.
- Y. Jararweh, A. Doulat, O. AlQudah, E. Ahmed, M. Al-Ayyoub, and E. Benkhelifa, "The future of mobile cloud computing: integrating cloudlets and mobile edge computing," in Proc. of 23rd Int. Conf. Telecommun. (ICT). IEEE, pp. 1-5, 2016.
- F. Bonomi, R. Milito, J. Zhu, and S. Addepalli, "Fog computing and its role in the internet of things," in Proc. of 1st Edition MCC Workshop Mobile Cloud Comput. ACM, pp. 13-16, 2012.
- G. I. Klas, "Fog computing and mobile edge cloud gain momentum open fog consortium, etsi mec and cloudlets," Google Scholar, 2015.
- A. Al-Shuwaili and O. Simeone, "Energy-efficient resource allocation for mobile edge computing-based augmented reality applications," IEEE Communication Letter, vol. 6, no. 3, pp. 398-401, 2017. https://doi.org/10.1109/LWC.2017.2696539
- X. Chen, L. Jiao, W. Li, and X. Fu, "Efficient multi-user computation offloading for mobile-edge cloud computing," IEEE/ACM Transaction on Networking, vol. 24, no. 5, pp. 2795-2808, 2015. https://doi.org/10.1109/TNET.2015.2487344
- X. Chen, "Decentralized computation offloading game for mobile cloud computing," IEEE Transaction on Parallel Distribution System, vol. 26, no. 4, pp. 974-983, 2014. https://doi.org/10.1109/TPDS.2014.2316834
- T. Q. Dinh, J. Tang, Q. D. La, and T. Q. Quek, "Adaptive computation scaling and task offloading in mobile edge computing," in Proc. of WCNC. IEEE, pp. 1-6, 2017.
- F. Wang, "Computation rate maximization for wireless powered mobile edge computing," in Proc of 23rd Asia-Pacific Conf. Commun. (APCC). IEEE, pp. 1-6, 2017.
- S. Bi and Y. J. Zhang, "Computation rate maximization for wireless powered mobile-edge computing with binary computation offloading," IEEE Transaction Wireless Communication, vol. 17, no. 6, pp. 4177-4190, 2018. https://doi.org/10.1109/twc.2018.2821664
- Y. Mao, J. Zhang, and K. B. Letaief, "Dynamic computation offloading for mobile-edge computing with energy harvesting devices," IEEE Journal on Selected Areas in Communication, vol. 34, no. 12, pp. 3590-3605, 2016. https://doi.org/10.1109/JSAC.2016.2611964
- X. Hu, K.-K. Wong, and K. Yang, "Wireless powered cooperation-assisted mobile edge computing," IEEE Transaction on Wireless Communication, vol. 17, no. 4, pp. 2375-2388, 2018. https://doi.org/10.1109/twc.2018.2794345
- F. Wang, J. Xu, X. Wang, and S. Cui, "Joint offloading and computing optimization in wireless powered mobile-edge computing systems," IEEE Transaction on Wireless Communication, vol. 17, no. 3, pp. 1784-1797, 2017. https://doi.org/10.1109/twc.2017.2785305
- Y. Liu, S. Wang, and F. Yang, "Poster Abstract: A multi-user computation offloading algorithm based on game theory in mobile cloud computing," in Proc. of IEEE/ACM Symp. Edge Comput. (SEC). IEEE, pp. 93-94, 2016.
- W. Zhang, Y. Wen, and D. O. Wu, "Collaborative task execution in mobile cloud computing under a stochastic wireless channel," IEEE Transaction on Wireless Communication, vol. 14, no. 1, pp. 81-93, 2014. https://doi.org/10.1109/TWC.2014.2331051
- S. E. Mahmoodi, K. Subbalakshmi, and V. Sagar, "Cloud offloading for multi-radio enabled mobile devices," in Proc. of IEEE Int. Conf. Commun. (ICC). IEEE, pp. 5473-5478, 2015.
- Y.-H. Kao, B. Krishnamachari, M.-R. Ra, and F. Bai, "Hermes: Latency optimal task assignment for resource-constrained mobile computing," IEEE Transaction on Mobile Computing, vol. 16, no. 11, pp. 3056-3069, 2017. https://doi.org/10.1109/TMC.2017.2679712
- S. Khalili and O. Simeone, "Inter-layer per-mobile optimization of cloud mobile computing: a message-passing approach," Transaction on Emerging Telecommunications Technology, vol. 27, no. 6, pp. 814-827, 2016. https://doi.org/10.1002/ett.3028
- S. E. Mahmoodi, R. Uma, and K. Subbalakshmi, "Optimal joint scheduling and cloud offloading for mobile applications," IEEE Transaction on Cloud Computing, 2016.
- P. Di Lorenzo, S. Barbarossa, and S. Sardellitti, "Joint optimization of radio resources and code partitioning in mobile edge computing," arXiv preprint arXiv:1307.3835, 2013.
- S. Cao, X. Tao, Y. Hou, and Q. Cui, "An energy-optimal offloading algorithm of mobile computing based on HetNets," in Proc. of 2015 International Conference on Connected Vehicles and Expo (ICCVE), Shenzhen, China, pp. 254-258, 2015.
- J. Kennedy and R. C. Eberhart, "A discrete binary version of the particle swarm algorithm," in Proc. of IEEE International conference on systems, man, and cybernetics. Computational cybernetics and simulation, Orlando, FL, USA, pp. 4104-4108, 1997.
- A. Bhattcharya and P. De, "Computation offloading from mobile devices: Can edge devices perform better than the cloud?," in Proc. of ARMS-CC. ACM, pp. 1-6, 2016.
- M. Safar, I. Ahmad, and A. Al-Yatama, "Energy-aware computation offloading in wearable computing," in Proc. of Int. Conf. Comput. Appl. IEEE, pp. 266-278, 2017.
- A. R. Jensen, M. Lauridsen, P. Mogensen, T. B. Sorensen, and P. Jensen, "Lte ue power consumption model: For system level energy and performance optimization," in Proc. of IEEE VTC Fall. IEEE, pp.1-5, 2012.
- G. Auer, O. Blume, V. Giannini, I. Godor, M. Imran, Y. Jading, E. Katranaras, M. Olsson, D. Sabella, P. Skillermark et al., "D2. 3: Energyefficiency analysis of the reference systems, areas of improvements andtarget breakdown," Earth, vol. 20, no. 10, 2010.
- S. Cui, A. J. Goldsmith, and A. Bahai, "Power estimation for Viterbi decoders," Wireless Systems Lab, Stanford Univ., Stanford, CA, USA, Tech. Rep., 2003.
- O. Munoz, A. Pascual-Iserte, and J. Vidal, "Optimization of radio and computational resources for energy efficiency in latency-constrained application offloading," IEEE Transaction on Vehicular Technology, vol. 64, no. 10, pp.4738-4755, 2014. https://doi.org/10.1109/TVT.2014.2372852
- S. Boyd and L. Vandenberghe, "Convex optimization," Cambridge university press, 2004.
- X. Cao, F. Wang, J. Xu, R. Zhang, and S. Cui, "Joint computation and communication cooperation for mobile edge computing," in Proc. of IEEE WiOpt. IEEE, pp. 1-6, 2018.
Cited by
- Collaborative Task Offloading Strategy of UAV Cluster Using Improved Genetic Algorithm in Mobile Edge Computing vol.2021, 2020, https://doi.org/10.1155/2021/3965689
- Client-driven animated GIF generation framework using an acoustic feature vol.80, pp.28, 2021, https://doi.org/10.1007/s11042-020-10236-6