Multi-user multi-objective computation offloading for medical image diagnosis
- Published
- Accepted
- Received
- Academic Editor
- Junaid Shuja
- Subject Areas
- Bioinformatics, Computer Architecture, Computer Networks and Communications, Distributed and Parallel Computing, Optimization Theory and Computation
- Keywords
- Computation offloading, Risk awareness, Multi-objective, Prospect theory, Distributed optimization, Exact potential game
- Copyright
- © 2023 Liu et al.
- Licence
- This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
- Cite this article
- 2023. Multi-user multi-objective computation offloading for medical image diagnosis. PeerJ Computer Science 9:e1239 https://doi.org/10.7717/peerj-cs.1239
Abstract
Computation offloading has effectively solved the problem of terminal devices computing resources limitation in hospitals by shifting the medical image diagnosis task to the edge servers for execution. Appropriate offloading strategies for diagnostic tasks are essential. However, the risk awareness of each user and the multiple expenses associated with processing tasks have been ignored in prior works. In this article, a multi-user multi-objective computation offloading for medical image diagnosis is proposed. First, the prospect theoretic utility function of each user is designed considering the delay, energy consumption, payment, and risk awareness. Second, the computation offloading problem including the above factors is defined as a distributed optimization problem, which with the goal of maximizing the utility of each user. The distributed optimization problem is then transformed into a non-cooperative game among the users. The exact potential game proves that the non-cooperative game has Nash equilibrium points. A low-complexity computation offloading algorithm based on best response dynamics finally is proposed. Detailed numerical experiments demonstrate the impact of different parameters and convergence in the algorithm on the utility function. The result shows that, compare with four benchmarks and four heuristic algorithms, the proposed algorithm in this article ensures a faster convergence speed and achieves only a 1.14% decrease in the utility value as the number of users increases.
Introduction
Medical imaging examinations are currently required for over 70% of clinical diagnostic behaviors in hospitals (Jayashree & Bhuvaneswaran, 1970; Maglogiannis et al., 2017; Teng, Kong & Wang, 2019). However, medical data grows unusually fast with the advancement of information technology in smart medicine. There will be a total of 40 trillion GB of medical data in 2020, with 85–90% of that coming from medical imaging, which exacerbates the burden of doctors’ imaging diagnosis work. Doctors hope to use intelligent image diagnostic models (IIDM) for accelerating image diagnosis (Zivkovic et al., 2022), but the hospital personal computer configuration is not high enough to meet the demand.
A new hospital service architecture known as the medical imaging cloud has emerged (Lakshmi et al., 2021). It sends medical images to the IIDM in the central cloud for processing by utilizing cloud computing, big data, the Internet of Things, digital imaging technology, and Internet technology. Despite the rapid availability of diagnostic results, there are still questions. For instance, deep learning-based IIDM will generate various parameters, which will lead to a significant increase in computation. In addition, the extremely long-distance image transfer between the central cloud and the hospital takes up a huge amount of network bandwidth, resulting in large delay, energy consumption and communication overhead.
To overcome the above shortcomings, several researchers have recently found that computation offloading (Shakarami, Shahidinejad & Ghobaei-Arani, 2020, 2021; Xu et al., 2020; Tong et al., 2020; uz Zaman et al., 2021, 2022b) is a promising technology to solve this dilemma. Computation offloading, as one of the key technologies for edge computing (uz Zaman et al., 2022a), refers to the technology by which resource-constrained terminal devices (TDs) offload part or all of the computing tasks to the edge server execution. It comprises an offloading strategy and resource allocation (Li et al., 2020c). This article focuses on the former. Specifically, the diagnosis tasks of the TDs are first uploaded and then processed on the edge servers. Finally, the corresponding terminal device receives the results (Zhang et al., 2019b).
Many researchers have conducted extensive research on computation offloading. There are the following major issues, however, with the existing offloading works: (1) the edge server will provide resources to the terminal devices without charge. However, in a real communication and computing environment, the cost of computing resources and wireless communication is unavoidable. Doctors cannot enjoy edge server services for free, but have to pay a fee. (2) It is not possible to effectively integrate real environmental concerns into their decision-making. Although performing diagnostic tasks on the edge nodes decreases latency, it might increase energy consumption and payment. Therefore, it is important to propose a trade-off offloading model between latency, energy, and cost. (3) Users hold a risk-neutral attitude. However, it has been argued recently that users are risk awareness when using edge server’s resources, especially in a resource-constrained environments (Vamvakas, Tsiropoulou & Papavassiliou, 2019a). Specifically, users can be classified into aggressive and conservative according to their behavior characteristics. For aggressive users, they will exhibit risk-seeking behavior, who want to offload diagnosis tasks to edge servers to avoid using resources on terminal devices, even though edge servers may not provide data processing services for all users. Another type of conservative user exhibits risk-aversion behavior, who prefers to process diagnosis tasks on terminal devices. The reason is that the computing resources of edge server will be overused when multiple users use it simultaneously.
To be closer to the real communication and computing environment, therefore, when making the offloading decision of each user in the medical image cloud scenario, multiple factors are considered. The factors include the user’s risk awareness and a set of innovative objectives: delay, energy consumption and payment. Our ultimate optimization goal is to maximize the prospect theoretic utility of each user. To achieve this goal, we propose a multi-user multi-objective computation offloading for medical image diagnosis. First, we design the user’s utility function based on the Prospect Theory principle by combining the multi-objective of delay, energy consumption, and payment, which simulates the risk awareness behavior of each user during diagnosis task offloading. Second, to maximize the utility, the computation offloading problem first is expressed as a distributed optimization problem, then is transformed into a non-cooperative game among the users. Third, we prove that this game has Nash equilibrium (NE) points based on exact potential game and propose a low complexity computation offloading algorithm based on best response dynamics (BRD-CO) to reach an NE point. Finally, we conduct detailed simulation experiments. The results show that the BRD-CO algorithm can guarantee that each user has a higher prospect theoretic utility and a faster convergence speed when compared with four benchmarks and four heuristic algorithms.
Therefore, according to the above, this article proposes a computation offloading method that employs the following two element problems as a guide for investigation:
The possibility of designing a more realistic optimization goal function based on user risk awareness and multiple objectives.
The possibility of further improving the convergence speed of the offloading algorithm in a distributed manner.
Based on the system model constructed, the framework designed and the experimental results, the main contributions of this work are summarized as follows.
We develop a more specific and detailed computation offloading model using the formal method. It more clearly reflects the execution process of the user’s diagnosis task on the edge server and terminal devices, respectively.
We achieve a more realistic optimization goal. The multi-user and multi-objective computation offloading method are closer to the real world, which not only reflects the risk attitude of each user but also trade-off for delay, energy consumption and payment.
We design a distributed offloading algorithm with a faster convergence speed. Each user wants more for computation and wireless communication resources during the execution of a diagnostic task, the computation offloading problem therefore is considered as a distributed optimization problem. We propose an optimal computation offloading algorithm based on best response dynamics, which requires only a few iterations to converge to the Nash equilibrium point.
We have achieved a higher prospect theoretic utility. We implement the proposed BRD-CO algorithm and conduct detailed studies. The experimental results show that the proposed algorithm has statistical superiority and provides a higher prospect theoretic utility.
The rest of this article is organized as follows. “Related work” presents the related work. “Computation offloading system model” illustrates the computation offloading model and discusses the delay, energy consumption and payment under different offloading modes. “Multi-user multi-objective computation offloading for medical image diagnosis” introduces a multi-user multi-objective computation offloading for medical image diagnosis. “Numerical results” designs the simulation experiment and presents the numerical results. Finally, a summary of our work and future plan is presented in “Conclusions”.
Related work
Massive medical image data is becoming more challenging to process and manipulate as the advancement of medical information (Zhang et al., 2017b). As a way of managing and procedure big data, cloud computing plays an important role (El-Seoud et al., 2017; Rahman, Khalil & Yi, 2019). Zhang et al. (2020) proposed a normal distribution splitting-based method for executing plenty of medical data parallel. On the other hand, we can use the parallel computing and data distribution functions of related systems, such as the MapReduce model and Hadoop model (Khezr & Navimipour, 2017; Mo, 2019; Duan, Edwards & Dwivedi, 2019). Based on the Hadoop, MapReduce and Spark, the researcher uses machine learning to predict and analyze the future complications of diabetic patients, which improved processing speed (Vineetha & Nandhana, 2022). In the framework of medical imaging cloud based on cloud computing, however, the distance between the central cloud and the hospital is so far that the transmission will consume a large amount of bandwidth and cause huge latency.
Recently, computation offloading has received more and more attention as one of the most promising solutions to this issue, and various offloading strategies have been proposed (Mao, Zhang & Letaief, 2016; Zhang et al., 2017a, 2019a; Guo, Li & Guan, 2019; Li et al., 2019b, 2020b, 2020a; Messous et al., 2019; Meng et al., 2019; Mitsis, Tsiropoulou & Papavassiliou, 2020; Zhu et al., 2020a, 2020b; Alioua et al., 2020; Tang & Wong, 2022; Wang et al., 2021; Chen & Liu, 2021). The differences between various computation offloading methods are shown in Table 1. There are currently only a few studies on accelerating the processing of medical image data by computation offloading, mostly focusing on areas such as the internet of vehicle, unmanned aerial vehicles, etc. On the grounds of the optimization goal, these strategies can be divided into four categories: reducing delay, reducing energy consumption (EC), balancing delay with energy consumption, and maximizing utility.
References | Utilized technique | Performance metrics |
Evaluation tools |
Case study | Advantages | Disadvantages |
---|---|---|---|---|---|---|
Meng et al. (2019) |
|
|
|
|
|
|
Mao, Zhang & Letaief (2016) |
|
|
|
|
|
|
Zhu et al. (2020a) |
|
|
|
|
|
|
Tang & Wong (2022) |
|
|
|
|
|
|
Wang et al. (2021) |
|
|
|
|
|
|
Zhang et al. (2019a) |
|
|
|
|
|
|
Chen & Liu (2021) |
|
|
|
|
|
|
Li et al. (2020b) |
|
|
|
|
|
|
Zhu et al. (2020b) |
|
|
|
|
|
|
Guo, Li & Guan (2019) |
|
|
|
|
|
|
Zhang et al. (2017a) |
|
|
|
|
|
|
Li et al. (2019b) |
|
|
|
|
|
|
Messous et al. (2019) |
|
|
|
|
|
|
Alioua et al. (2020) |
|
|
|
|
|
|
Li et al. (2020a) |
|
|
|
|
|
|
Mitsis, Tsiropoulou & Papavassiliou (2020) |
|
|
|
|
|
|
Mao, Zhang & Letaief (2016), Meng et al. (2019), Zhu et al. (2020a), Tang & Wong (2022) are offloading strategies to reduce the delay. For instance, Mao, Zhang & Letaief (2016) proposed a dynamic offloading method based on Lyapunov optimization, considering the execution latency and task failure, which can decrease the task time by 64%. However, these offloading strategies are only designed to minimize the overall delay, without considering the potential energy consumption.
Zhang et al. (2019a), Li et al. (2020b), Wang et al. (2021), Chen & Liu (2021) are offloading strategies to reduce energy consumption. For instance, Wang et al. (2021) proposed a trajectory control algorithm based on convex optimization and deep reinforcement learning by combining the motion trajectory, user association, and resource allocation of UAVs. Similar to the previous optimization goal, these offloading strategies are only effective in reducing the overall energy consumption of the task. However, in some systems, users prefer to achieve relative stability between delay and energy consumption.
Zhang et al. (2017a), Guo, Li & Guan (2019), Li et al. (2019b), Zhu et al. (2020b) are offloading strategies to balance delay and energy consumption. For instance, Zhang et al. (2017a) introduced predation behavior, swarm behavior and following behavior into the artificial fish swarm algorithm, which saves 30% energy consumption. While these offloading strategies achieve a tradeoff between latency and energy consumption, they may not be applicable to all systems. The reason for this is that each system has different performance requirements, not all of which are latency and energy consumption.
Messous et al. (2019), Mitsis, Tsiropoulou & Papavassiliou (2020), Alioua et al. (2020), Li et al. (2020a) are offloading strategies to maximizing utility. For instance, Messous et al. (2019) used game-theoretical to reach a balance of energy consumption, delay and payment. Similarly, Alioua et al. (2020) also proposed a sequential game-based computation offloading strategy. Li et al. (2020a) proposed an algorithm jointly optimizing the delay and payment for task offloading. Mitsis, Tsiropoulou & Papavassiliou (2020) proposed a resource-based pricing and user’s risk awareness computation offloading scheme. The above-mentioned offloading strategies have been widely concerned because they can design different utility functions according to different scenarios, and create an appropriate offloading strategy to meet the needs of users.
In the view of prior works, few studies conjointly consider risk awareness, delay, energy consumption, and payment. Most of which focus on two or three aspects, and assumes that the risk-neutral behavior of the users in the process of task offloading. To simulate the resource consumption in the real-world environment, in this article, we propose a computation offloading model for maximizing the prospect theoretic utility of each user, which jointly considers: (1) a clearer formal description for the computation offloading model; (2) more realistic optimization goals; (3) a distributed offloading algorithm with a faster convergence speed; (4) higher prospect theoretic utility. We also conduct experiments to evaluate the BRD-CO algorithm under various parameters.
Computation offloading system model
The scenario in this article is a medical image diagnosis in a medical image cloud. In this section, we construct a computation offloading system model and introduce three offloading modes.
Notation description
For readability, Table 2 summarizes the notation used in this article.
Notation | Description | Notation | Description | Notation | Description |
---|---|---|---|---|---|
The finite set of terminal devices | The device used by the i-th user | The finite set of diagnosis tasks | |||
The computation task of | The finite set of offloading proportions | The offloading proportion of | |||
Task computational complexity | The expected profits | The user’s anticipated profit | |||
The user’s actual profit | The loss aversion coefficient | Delay weight | |||
Energy consumption weight | Payment weight | Payment factor | |||
The failure probability of the edge server | Gain attitude | Loss attitude | |||
The finite set of local computing parameters | The finite set of the computational capability | The computational capability of | |||
The finite set of the energy coefficient | The consumed energy per CPU cycle of | The finite set of the computation delay locally | |||
The computation delay required by to process locally | The finite set of the computation energy consumption locally | The computation energy consumption required by to process locally. | |||
The finite set of full offloading parameters | The computational capability of the edge server | The energy coefficient | |||
The finite set of computation delay pricing | The computation delay pricing of | The finite set of transmission power | |||
The transmission power between and the edge server | The finite set of the transmission rate | Transmission rate between and the edge server, | |||
The finite set of the total delay on the edge server | , | The risk attitude coefficient | The total delay required by to process on the edge server | ||
The finite set of the computation delay on the edge server | The computation delay required by to process on the edge server | The finite set of the transmission delay on the edge server | |||
The transmission delay required by to process on the edge server | The finite set of the total energy consumption on the edge server | The total energy consumption required by to process on the edge server. | |||
The finite set of the computation energy consumption on the edge server. | The computation energy consumption required by to process on the edge server | The finite set of the transmission energy consumption on the edge server | |||
The transmission energy consumption required by to process on the edge server | The finite set of the payment | The finite set of the computation delay on the edge server | |||
The computation delay cost required by to process on the edge server | The finite set of partial offloading parameters | The finite set of the total delay in partial offloading | |||
Denotes the delay required by to process in partial offloading, | The finite set of the total energy consumption in partial offloading | The energy consumption required by to process in partial offloading | |||
The finite set of the total delay in partial offloading |
The computation delay cost required by to process in partial offloading |
System model description
In the concerned scenario, we consider a medical image cloud that includes an edge server and multiple terminal devices. The edge server provides storage and computing services for users, solving the problem of limited resources for terminal devices. The terminal devices are used by doctors, which include desktops, laptops, tablets and super beans. Each terminal device is equipped with computing resources for processing diagnosis tasks.
As the computing resources of terminal devices are limited, they cannot meet the needs of massive medical image diagnosis tasks. Therefore, users will offload part or all diagnosis tasks to the edge server. Such behavior of users carries a risk-aversion or risk-seeking attitude. The edge server typically charges a payment to share their resources. In addition, terminal devices and the edge server cause delay and energy consumption when performing diagnosis tasks. To clearly explain the offloading process of the image diagnosis task in the medical image cloud, it is formalized as follows:
Definition 1. Computation offloading system model.
The computation offloading system model is an eight-tuple, the details of which are shown in the Supplemental Information.
Offloading modes
Each terminal device has one or more medical image diagnosis tasks to perform. As shown in Fig. 1, there are three offloading modes for the diagnosis tasks: local computing, full offloading and partial offloading. Each offloading mode can be conceived as three-stages, including the sending, processing and feedback steps. First, the part or full diagnosis tasks are sent from to the edge server. Second, the diagnosis task offloaded is processed on the edge server. Third, the processed results are feedback to . Next, the computing methods of three objectives in different modes are as follows.
Local computing
In the local computing mode, as illustrated in Fig. 1, users execute the diagnosis task [bits] only using the computing resources of terminal devices, where the offloading proportion . For terminal device , the local computation delay [s] of processing can be given by
(1)
Besides the required computation delay, each diagnosis task also consumes some computation energy. Therefore, the local computation energy consumption [J] required by to process can be given by
(2)
Full computing
In the full offloading mode, as illustrated in Fig. 2, the diagnosis task had to be performed completely on the edge server, where the offloading proportion . Therefore, the transmission delay [s] required by to process on the edge server via the uplink channel can be given by
(3)
The transmission energy consumption [J] required by to process on the edge server can be given by
(4)
Next, the edge server will use some of the computing resources to perform . Therefore, the computation delay [s] required by to process on the edge server can be given by
(5)
Meanwhile, computation energy consumption is also generated. Therefore, the computation energy consumption [J] required by to process on the edge server can be given by
(6)
After the diagnosis task is completed, the results will be sent back to terminal devices via the downlink channel. However, resembling many studies (Xian, Lu & Li, 2007; Wang et al., 2017; Cui et al., 2017; Rudenko et al., 1998), we ignore the downlink transmission delay because the results are insufficient compared to the original image data.
In summary, the total delay [s] required by to process on the edge server can be given by
(7)
The total energy consumption [J] required by to process on the edge server can be given by
(8)
We assume that the user has to pay a fee for the edge server based on the computation delay pricing [$/s] (see “Prospect Theoretic Utility”) and the computation delay . Therefore, the payment [$] required by to process on the edge server can be given by
(9)
To simplify the model, we assume that the configuration and transmission setting are the same for each terminal device in this article (i.e., and ).
Partial computing
In the partial offloading mode, as illustrated in Fig. 3, the diagnosis task is divided into two subtasks, where the offloading proportion . Subtask performed on the edge server while subtask will be executed on . Therefore, the total delay , total energy consumption and payment required by to process in partial offloading can be given by
(10)
It can be clearly seen that when the computation delay pricing is constant, the greater amount of subtask offload, the more payment users will pay.
Multi-user multi-objective computation offloading for medical image diagnosis
In this section, we propose a multi-user multi-objective computation offloading for medical image diagnosis. First, the proposed framework of this system model is presented in detail. We then construct the user’s prospect theoretic utility function, formulate the offloading problem as a non-cooperative game among users, and design an algorithm to solve the problem.
Proposed framework overview
In the above computation offloading system model, users are not always risk-neutral when deciding where to process diagnosis tasks. The reason is that the different offloading modes will yield different profits for the users. Faced with possible future gains and losses, users can hardly be in a completely neutral attitude, but will indicate different risk attitudes depending on situation.
Therefore, considering the delay, energy consumption, payment and user’s risk awareness behavioral characteristics required to complete the diagnosis tasks, we propose a multi-user multi-objective computation offloading for medical image diagnosis. The proposed framework consists of two parts, as shown in Fig. 2. (1) Problem formulation: computation offloading problem involving delay, energy consumption, payment and risk awareness behavioral characteristics, whose optimization goal is to maximize the user’s utility based on Prospect Theory. This problem is regarded as a DO problem and then formulated as an NCG among users. (2) Problem solution: the complete proof of the existence of the NE is provided by the exact potential game (EPG). Then, we propose a low-complexity computation offloading algorithm based on best response dynamics (BRD-CO), to determine the optimal data offloaded in a distributed manner for each user.
Problem formulation
Computation offloading problem involving risk awareness and multi-objective is regarded as a distributed optimization problem. Its optimization goal is to maximize the user’s prospect theoretic utility. Then, it is formulated as a non-cooperative game among users and solved by exact potential game.
Prospect theoretic utility
When users make an offloading strategy, we will analyze their risk-aware behavior using the prospect theory (PT). The prospect theory was first proposed after revising the expected utility theory based on absorbing “Allais Paradox” in 1979 (Kahneman & Tversky, 1988). The theory combines psychology and behavioral science, explicitly states that humans exhibit “loss aversion” when deciding. When faced with gains, users exhibit an attitude of risk-aversion. When faced with losses, users exhibit a risk-seeking attitude and have the principle of being more sensitive to losses than gains (Wu & Gonzalez, 1996).
Specifically, prospect theory simplifies the results by establishing appropriate reference points and a preliminary analysis of various outcomes during the editing phase. Then, the decision with the highest PTU is selected in the evaluation phase by figuring the results of the previous phase through a value function (VF) and a weight function (WF).
Each user offloads part or all of the diagnosis task to the edge server, as shown in Fig. 3. We then calculate the value of the edge server in different states based on the VF. When all users have offloaded, we calculate the failure probability of the edge server. Then, the probability is modified to a weight according to the WF. Finally, we obtain the user’s PTU by multiplying the corresponding value and weight. The specific definitions of VF and WF are as follows.
Value function
The value function mainly reflects the subjective value of users, following the principle of PT (Vamvakas, Tsiropoulou & Papavassiliou, 2019b), which can be given by
(11)
Inspired by Tram, Tham & Niyato (2014), Zhou, Tham & Motani (2017), Li et al. (2019a), denotes the reference point, expressing the user’s anticipated profit by fully processing diagnosis task at . represents the user’s actual profit after offloading part or all of the diagnosis task to the edge server, and is given by (12) below. and represent risk attitude coefficient, which are . As the and increase, the risk taken by the user becomes greater. , the user is risk-neutral. is the loss aversion coefficient. indicates that users are more stimulated by losses instead of than gains. Moreover, the user can adjust , and in different environments. For simplicity, we attempt to assume in this article.
, and denote the multi-objective weight coefficient, i.e., delay weight, energy consumption weight and payment weight. We map these different measures into the same dimension, where , , . For delay-sensitive tasks, is larger than and . For energy-sensitive tasks and payment-sensitive tasks, and are relatively large.
(12)
Given the weak computing capacity of a terminal device, it cannot meet the computing needs of a massive medical image. As a common resource, the edge server can provide services for all users. Every user can enjoy edge server services, but the computing resources of the edge server are limited. There will be serious negative effects when the resources use exceeds the boundary. Here, we mainly divide it into two situations.
Situation 1: edge server survives. There may be some signal interference or channel congestion, resulting in reduced transmission efficiency, but the edge server remains capable of diagnosis tasks of terminal devices.
Situation 2: edge server failures. Excessive competition for computing resources on edge server, terminal devices will no longer be able to enjoy services once edge server is shutdown.
The first branch of (12) denotes the actual profit of the user performing tasks entirely on the terminal device. The second branch of (12) corresponds to situation 1, where the user’s actual profit depends primarily on the delay, energy consumption and payment after executing all the images. Quite the opposite, the third branch of (12) corresponds to situation 2, where the user’s actual benefit is determined by the energy consumption and payment. The reason is that the edge server is shutdown and is no longer able to get a delay gain after processing the image.
Therefore, the living state of the edge server directly affects the user’s actual profits. In situation 1, the value function of the user should be determined by the first branch of (11) and the second branch of (12), which is defined as
(13)
In situation 2, the value function of the user should be resolute by the second branch of (11) and the third branch of (12), which is defined as
(14)
According to the math characteristics of the value function, (13) must be a positive constant, and (14) must be a negative constant. Thus, we can determine the boundaries of the computing delay pricing imposed by the edge server on the user, which can be given by
(15)
For simplicity, , so , where payment factor .
Weight function
The weight function reflects the degree of perception of probability. Users have different risk behaviors towards different failure probabilities of the edge server during the diagnosis task offloading (i.e., Gains and losses). The failure probability of the edge server is directly related to the size of the processed data. The reason is that the larger the offloaded amount, the higher the computing demand for the terminal devices. This will lead to a greater failure probability of the edge server. Inspire by Mitsis, Tsiropoulou & Papavassiliou (2020), the failure probability of the edge server can be given by
(16) where represents the offload image data of the user on the edge server. is a positive constant calibrating the sigmoidal curve based on the computing capabilities of the edge server (Mitsis, Tsiropoulou & Papavassiliou, 2020). The failure probability of the edge server, , is a continuous, strictly increasing, convex, and twice differentiable function (Mitsis, Tsiropoulou & Papavassiliou, 2020).
Following the principle of PT, we convert the probability function into the weight function , which is defined as
(17)
The parameter denote the risk attitude of the user to gains and losses in making an offloading strategy. is the increment function of . When is a small probability event approaching 0, users show risk-seeking attitude (i.e., ). For the events with medium probability and high probability, the users show the attitude of risk aversion (i.e., ). In other words, low-probability events tend to be overestimated, and the converse holds (Monderer & Shapley, 1996).
Combining (11)–(17), following the principle of PT, the user’s prospect theoretic utility function comprises the value function and the weight function, which is defined as follows
(18) where denotes the weight of the edge server survives (i.e., Gains). represents the weight of the edge server fails (i.e., Losses). The definition is as follows
(19)
(20)
Problem model
To maximize the prospect theoretic utility of each user, the computation offloading problem for medical image diagnosis task involving risk awareness and multi-objective (i.e., delay, energy consumption and payment) is, therefore, represented as a distributed optimization problem as follows
(21) where is the amount of image data offloaded by the rest of the terminal devices except for the terminal device . The above problem is defined as a non-cooperative game among users as follows
(22) where is the finite set of the user terminal devices, is the offloading strategies space of , and reflects the prospect theoretic utility of the user i. The solution of for the user’s optimal computation offloading strategy , the meaning of which is that PTU is greatest when the amount of data offloaded is .
Definition 2. Nash equilibrium.
An image data offloading vector in the strategy space is a Nash Equilibrium point if for every user i the following condition holds true
(23)
The meaning of the Nash equilibrium point is that, no player (users in our problem) can further increase the cost (user’s prospect theoretic utility in our problem) by unilaterally changing his strategy while the other player’s strategy (computation offloading strategy in our problem) remains unchanged.
Problem solution
In this section, we first prove the existence of NE points for the NCG by EPG. Then, a computation offloading algorithm based on best response dynamics is proposed to solve the problem. Finally, we discuss the time complexity of the proposed algorithm.
The existence of NE point
To prove has at least one NE point, which means as a solution to maximize the distributed optimization problem, the exact potential game is adopted. The main reason for this design is that not all NCGs have an NE point and can reach algorithmic convergence. An exact potential game with a limited set of strategies converges to at least one NE point, regardless of the starting point.
Definition 3. Exact potential game. The is an EPG if there is an exact potential function that for all satisfies the following conditions
(24)
Theorem 1. The is an exact potential game and has at least one Nash equilibrium point . (Due to space limitation, the proof of theorem 1 is shown in the Supplemental Information).
Computation offloading algorithm based on best response dynamics
Given that we have already proven that the belongs to the class of EPG as stated above, and exists at least one NE point. In an exact potential game, the NE point can always be reached after a finite number of iterations, which is called the finitely increasing property (Yang et al., 2020). Therefore, the best response dynamics is adopted to determine each user’s optimal computation offloading strategy (i.e., converged to a NE point) in a distributed manner through a finite number of iterations, when the computation offloading strategy of other users is determined (Topkis, 1998; Milgrom & Roberts, 1990), as follows
(25)
From the above discussion, we propose a low-complexity computation offloading algorithm based on best response dynamics (BRD-CO) to determine each user’s computation offloading strategy. To more clearly describe the workflow of the BRD-CO algorithm, described briefly in Table 3, the algorithm follows a pseudo-code. The BRD-CO algorithm comprises three parts: the first part is a line 1–4, the initial definition of parameters, including the number of iterations, the user i’s computation offloading strategy and the convergence of the algorithm. The second part is lines 9–14, which calculates the user i’s prospect theoretic utility. In each iteration, first, we calculate the delay, energy consumption and payment in three offloading modes via (1)–(10). Second, using (13) and (14), the user i’s value in the survival and failure state of the edge server can be obtained. Then, we calculate the failure probability of the edge server via (16) and use (19) and (20) convert probability to weight. Finally, taking advantage of (18), we get the user prospect theoretic utility. The third part is 16–20 lines, which determine the user optimal offloading strategy. We calculate the user i’s offload strategy via (25) each time. If two adjacent times strategies are the same, the strategy is called the optimal offload strategy for the user i.
Algorithm 1: BRD-CO algorithm | |
---|---|
Input: all paraments of model, , , , , , , , , , , , , , | |
Output: optimal computation offloading strategy, | |
1. | // Initialization Paraments |
2. | //iterations |
3. | //The initial amount of image data offloaded for the user i |
4. | convergence false //Whether the algorithm converges |
5. | While convergence == false do |
6. | |
7. | While do |
8. | // computer prospect theoretic utility |
9. | computer delay, energy consumption and payment via (1)–(10) |
10. | computer via (13) |
11. | computer via (14) |
12. | obtain via (16) |
13. | convert (16) into weight and and via (19) and (20) |
14. | computer via (18) |
15. | // determine the optimal strategy |
16. | user i determines is based on via (25) |
17. | End while |
18. | If then |
19. | convergence true |
20. | End if |
21. | End while |
Time complexity of BRD-CO algorithm
In this section, the time complexity of the proposed BRD-CO algorithms is discussed. From the above pseudo-code analysis, it can be seen that the solution process of the BRD-CO algorithm is iterative, and its time complexity is mainly determined by three factors: the number of iterations, the number of users, and the complexity of the utility function. Specifically, assume that the number of iterations required for Algorithm 1 is the complexity of (25) is, the number of users is. In each iteration, the formula (25) is calculated for all users. In addition, the difference between the current and previous offloading amounts is compared. If the difference is within the error range, the convergence state is adjusted to true; otherwise, it is adjusted to false. Therefore, the time complexity of the BRD-CO algorithm is O(F ite n).
Numerical results
In this section, we introduce the simulation setting to build . The parameter influence analysis then is exhibited from five aspects. Next, we discuss the convergence of the algorithm. Finally, we compare the proposed BRD-CO algorithm with four benchmarks and four heuristic methods.
Simulation setup
To evaluate the parameter influence and convergence of the algorithm, we use PyCharm as the development tool for Python IDE. The performed simulations were executed on an Intel® Xeon (R) Silver 4114 CPU @ 2.20 GHz 40 with 128 GB RAM.
The contents of the simulation are as follows: Suppose that 25 users simultaneously offload part of the annotation tasks to the edge server. The data used in the experiment came from the dynamic panoramic PET data set of Henan Provincial People’s Hospital. Each user’s prospect theoretic utility is calculated via (18). Using the BRD-CO algorithm to explore the optimal offloading strategy for each user to maximize their PTU. Inspired by Apostolopoulos, Tsiropoulou & Papavassiliou (2020), the main parameters are given in Table 4.
Parameters | Value | Parameters | Value | Parameters | Value |
---|---|---|---|---|---|
bits | J/CPU-cycles | 1,000 CPU-cycles/bit | |||
0.5 | 1.2 | 0.2 | |||
5 $/bit | J/CPU-cycles | CPU-cycles/s | |||
CPU-cycles/s | dbm | 0.1 bits/s | |||
1 | 0.001 | 0.1 | |||
0.61 | 0.69 |
Parameter influence analysis
In “Impact of Computing Delay Pricing”, we consider the number of users (denoted by ‘N’) is set to 25 and parameter values as indicated in Table 4. However, in the rest of the analysis, a wide range of computing delay pricing , number of users, multi-objective weight coefficients ( , and ) and prospect theoretic parameters (risk attitude , gain attitude , loss aversion and loss attitude ) are considered.
Impact of computing delay pricing
In this section, we discuss the impact of the computing delay pricing , imposed by the edge server on the user’s computation offloaded strategy, where the payment factor is from 0.001 to 0.9. The simulation results are shown in Fig. 4. It can see that as increases, the average prospect theoretic utility of users gradually increases to its maximum after that slowly decreases (Fig. 4A). Specifically, when the is small (i.e., = 0.001), the edge server will charge a lower and users will offload a large amount of image data (Fig. 4B), which results in lower delay (Fig. 4C) and lower payment (Fig. 4E), higher energy consumption (Fig. 4D), lower prospect theoretic utility (Fig. 4A). But the edge server will bear huge computing pressure, which causes a higher (Fig. 4F). When further increases, the edge server will charge a higher . Users are not willing to use the computing resources of the edge server, the delay and payment will gradually increase, the energy consumption and will reduce, a lower prospect theoretic utility again. Therefore, we need to balance the computing delay pricing to maintain the user’s high-quality experience. In addition, Figs. 4G–4I indicate the joint distribution between average utility and average delay, average energy consumption and average cost, respectively. It is clear from the regression line that the average utility shows a decreasing, increasing and decreasing trend with the increase of the three, respectively.
Impact of the number of users
In this section, we discuss the impact of the number of users on the user’s computation offloading strategy, where the number of users is from 1 to 100. The simulation results are shown in Fig. 5. When the N is small (i.e., N = 1, 2, 5), the is very low (Fig. 5F) because the computing resources on the edge server are far greater than the needs of users. Users are inclined to offload a large amount of image data to the edge server (Fig. 5B), resulting in lower delay (Fig. 5C), higher payment (Fig. 5E), higher energy consumption (Fig. 5D), and higher prospect theoretic utility (Fig. 5A). As N further increases, the edge server is under more and more computing pressure. Users tend to offload a small amount of image data to the edge server, while the remaining image process on the local device, which results in higher delay, lower payment, lower energy consumption, and lower prospect theoretic utility. Similarly, Figs. 5G–5I indicate the joint distribution between average utility and average delay, average energy consumption and average cost, respectively. It is clear from the regression line that the average utility shows a decreasing, increasing and increasing trend with the increase of the three, respectively.
The relationship between computing delay pricing and number of users
In this section, we discuss the relationship between computation delay pricing and the number of users N. The simulation results are shown in Fig. 6.
Especially, in Figs. 6A–6C, the x-axis shows the payment factor , and the y-axis shows the average PTU under N = 25, 55 and 95, respectively. It is noted that the average PTU of users gradually increases to its maximum with the increase of after that slowly decreases. Moreover, the payment factor is different when the maximum PTU is reached under different user’s numbers. To explore the relationship between the two, we performed the following experiments. Figure 6D shows the computation delay pricing corresponding to the maximum utility of a different number of users. When the number of users is small (i.e., 1, 2), the computing pressure on the edge server is very small, so it will impose a smaller payment factor , i.e., a lower computation delay pricing . As the number of users further increases, the edge server is under more and more computing pressure. To reduce the possibility of failure, the edge server will control the number of users by increasing . Since that N = 85, there is no change in both convergence speed and convergence result.
Impact of the multi-objective weight coefficients
In this section, we discuss the impact of the multi-objective weight coefficients on the user’s computation offloaded strategy, where multi-objective weight coefficients . The simulation results are shown in Fig. 7.
In the subfigure of Figs. 7A and 7B, the x-axis shows the delay weight , and the y-axis shows the user’s average offloaded image data amount and failure probability of the edge server in each , where , . As the increases from 0.5 to 1, users are inclined to offload an enormous amount of image data (Fig. 7A). The reason is that the scenario set in this article is a delay-sensitive task, and users choose a larger rather than smaller . This will bring tremendous pressure to the edge server and increase the failure probability (Fig. 7B).
In the subfigure of Figs. 7C and 7D, the x-axis shows the energy consumption weight , and the y-axis shows the user’s average offloaded image data amount and failure probability of the edge server in different , where , . The reason is that users are insensitive to energy consumption for delay-sensitive tasks. As the increases from 0.001 to 0.1, a fewer images will be offloaded (Fig. 7C), which reduces the failure probability (Fig. 7D).
In the subfigure of Figs. 7E and 7F, the x-axis shows the payment weight , and the y-axis shows the user’s average offloaded image data amount and failure probability of the edge server in different , where , . For the same reason, users are also relatively insensitive to payments. As the increases from 0.001 to 0.1, a little image will be offloaded (Fig. 7E) and the failure probability of the edge server also will reduce (Fig. 7F).
Impact of the prospect theoretic parameters
In this section, we discuss the impact of the prospect theoretic parameters, including risk attitude , gain attitude , loss aversion and loss attitude on the user’s offloading strategy. The simulation results are shown in Fig. 8.
In the subfigure of Figs. 8A and 8B, the x-axis shows the , and the y-axis shows the user’s average offloaded image data amount and failure probability of the edge server in each . As the risk attitude increases from 0 to1, users are inclined to offload a larger amount of image data (Fig. 8A). The reason is that they will choose larger gains, not smaller gains. As the increases, the failure probability of the edge server will also increase (Fig. 8B). For the same reason, As the gain attitude increases from 0.1 to 1, users will have a larger average offloaded image data amount (Fig. 8C) and larger failure probability of the edge server (Fig. 8D).
In the subfigure of Figs. 8E and 8F, the x-axis shows the , and the y-axis shows the user’s average offloaded image data amount and failure probability of the edge server in each . As the loss aversion increases from 0.1 to 2, they offload fewer images to the edge server (Fig. 8E). The reason is that the and user’s loss aversion are positively correlated. This will lead to the failure probability of the edge server decreasing (Fig. 8F). For the same reason, As the loss attitude increases from 0 to 1, the user will have a lower average offloaded image data amount (Fig. 8G) and larger failure probability of the edge server (Fig. 8H).
Convergence analysis
We evaluate the convergence of the BRD-CO algorithm. The simulation results are shown in Fig. 9. In the subfigure, the x-axis shows the number of iterations, and the y-axis shows each user’s offloaded image data amount, PTU, delay, energy consumption, payment and failure probability of the edge server.
There are two lines on the above and below of the boxes, which represent the maximum and minimum offloaded amounts. The red solid line is the median of the amount of offloaded image data, representing the average level, and the green dotted line represents the average level of offloaded image data after using the BRD-CO algorithm. Figures 9A–9E shows the convergence of each user’s offloaded image data amount, prospect theoretic utility, delay, energy consumption and payment after the image is offloaded to the edge server. The results show that the BRD-CO algorithm converged faster with fewer iterations.
Figure 10 shows the user’s average prospect theoretic utility, average delay, average energy consumption and average payment with the increase of iteration number. Specifically, each user will offload a large image on the edge server instead of the terminal device when ite = 1 (Figs. 9A and 10A). This behavior is highly likely to trigger the overuse of computing resources, which makes the increase dramatically (Fig. 9F). Therefore, the edge servers will curb this occurrence by charging users for higher energy consumption (Fig. 10C) and higher payment (Fig. 10D). Surprisingly, the algorithm will reach the lowest delay (Fig. 10B) because of the high computing power of the edge server and performing massive diagnosis tasks.
Considering the failure probability of the edge server, as ite increases from 2 to 5, the convergence speed of the BRD-CO algorithm is increase. It will lead to the rapid decrease of image data offloaded amount (Fig. 9A), prospect theoretic utility value (Fig. 10A), energy consumption (Fig. 10C) and payment (Fig. 10D) and the rapid increase of delay (Fig. 10B). As it further increases from 5 to 14, there is no change in both convergence speed and convergence result.
Method comparison
In this session, to evaluate the proposed BRD-CO method, we provide a comparative study between the proposed method with the following four benchmarks and four heuristic methods.
(1) Local computing (denoted by L. Comp.): all tasks are executed on the user terminal without offloading.
(2) Full offloading (denoted by F. Offl.): all tasks are executed on the edge server.
(3) Random offloading (denoted by R. Offl.): each task is randomly offloaded to the user terminal or edge server.
(4) Greedy offloading (denoted by G. Offl.): find the best offloading location for each task by selecting the current optimal solution each time.
(5) Particle swarm optimization-based offloading (denoted by PSO. Offl.) (Yuan et al., 2022): PSO simulates the foraging behavior of a flock of birds, using collaboration and information sharing among individuals in the flock to find the best decision to determine the offloading position of each task.
(6) Differential evolution-based offloading (denoted by DE. Offl.) (Hussain & Beg, 2021): DE simulates biological evolution by iterating repeatedly so that those individuals that are adapted to the environment are retained and the offloading position of each task is determined.
(7) Simulate anneal-based offloading (denoted by SA. Offl.) (Li, 2021): SA algorithm draws on the similarities, which exist between the annealing process of solids in statistical physics and general combinatorial optimization problems, to find the execution position of each task.
(8) Ant colony optimization-based offloading (denoted by ACO. Offl.) (Lin, Pankaj & Wang, 2018): each ant in the ACO algorithm uses pheromones to search simultaneously and independently at multiple points in the problem space, eventually finding the offloading position for each task.
Since the optimization goal of this article is to maximize the user’s prospect theoretic utility, we choose average utility as the performance metric. To eliminate the stochastic introduced by the heuristic algorithms, we conduct 50 runs and used the mean and standard deviation of PTU value as the final result. Table 5 and Table 6 in the Supplemental Information show the results of various heuristic algorithms for different payment factor and different numbers of users, respectively.
In Fig. 11A, as the payment factor increases, the proposed BRD-CO algorithm can always maintain a higher average utility when compared with the benchmark methods. When the is small (i.e., = 0.001), the maximum average utility can be achieved by most methods, but it is maximum in our method. As its future increases from 0.001 to 0.9, the average utility of our proposed BRD-CO algorithm decreases by 1.14%, while those of the benchmark methods (expect L. Offl.) decrease by at least 16.4%. This implies that as the computing delay pricing increases, the average utility of the proposed algorithm decreases less dramatically than other methods. The reason is that when the payment factor is larger, the users are inclined to offload fewer image, resulting in the average utility decreasing slowly.
In Fig. 11B, the proposed BRD-CO algorithm always achieves a higher average utility than the benchmark methods, especially when the number of users is small. As the number of users increases, the average utility of each algorithm (expect L. Offl.) decreases and gradually converges. The reason is that when the number of users is larger, the potential pressure at the edge servers will be increased. When the number of users is large enough, each user has very little offload. In addition, in this article, a marginal decrease of less than 0.005 in average utility is called convergence. From Fig. 11B, we can find that only the BRD-CO algorithm can converge when the number of users is 45. For G. Offl., R. Offl., F. Offl., PSO. Offl., DE. Offl., SA. Offl., and ACO. Offl., convergence is achieved at 50, 85, 55, 85, 85, 75, and 100, respectively (expect L. Offl.). When the BRD-CO reaches convergence, the average utility is 0.138, which is greater than G. Offl. (0.115), L. Offl. (0), R. Offl. (−2.052), F. Offl. (−2.523), PSO. Offl. (−2.500), DE. Offl. (−2.551), SA. Offl (−2.39) and ACO (−2.265). Therefore, compared with other methods, our proposed BRD-CO algorithm has a faster convergence speed and higher average utility.
Statistical test is an effective way to evaluate the performance of algorithms. In this article, the Wilcoxon rank sum test (Derrac et al., 2011) is adopted as a non-parametric statistical test that returns a P-value that verifies the significant level difference between the two algorithms. It is worth noting that an algorithm is statistically different when the P-value is less than 0.05. The P values obtained from formula (21) under different payment factor and different number of users are shown in Table 7 of the Supplemental Information. By evaluating the comparison between BRD-CO and the other eight algorithms, it is clearly understood that only one of the 16 P-values exceeds 0.05, which reflects the statistical superiority of BRD-CO.
Conclusions
In this article, we propose a multi-user multi-objective computation offloading for medical image diagnosis, which can play a significant role in the medical image cloud. Prior computation offloading strategies ignored payment required to perform computation tasks and a user’s risk awareness. To reflect the real communication and computing environment, we consider a more realistic optimization of multi-objective. Specifically, to maximize the prospect theoretic utility of each user by considering delay, energy consumption, payment and user’s risk awareness, we design a low-complexity BRD-CO algorithm. The algorithm can quickly converge to NE point and obtain an optimal computation offloading strategy for each user in a distributed manner. The parameter influence analysis of the BRD-CO algorithm is verified by five aspects. The simulation results show that when compared with four benchmarks and four heuristic algorithms, our proposed BRD-CO algorithm can guarantee a higher user’s prospect theoretic utility and a faster convergence speed. The benefit is especially significant when the diagnosis tasks are delay-sensitive and the resources of terminal devices are limited.
It is worth noting that the medical tasks studied in this article are coarse-grained, but some medical tasks can be more fine-grained. Therefore, our future research work focuses on task dependencies. For diagnostic tasks based on radiomics model, the association relationships between sub-modules (i.e., subtasks) within the model have a large impact on the task offloading problem. Therefore, we intend to use the recurrent neural network to model the dependencies between subtasks. Meanwhile, the high-quality task offloading strategy is learned by continuous ‘trial and error’ through deep reinforcement learning.