Professional Documents
Culture Documents
Noma 1
Noma 1
fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
IEEE TRANSACTIONS ON CLOUD COMPUTING 1
Abstract—The 5G network presents a new horizon with tremendous opportunities for future generation wireless networks. Mobile edge
computing (MEC), non-orthogonal multiple access (NOMA), and network slicing (NS) are some of the key enablers for 5G. MEC reduces
the latency to a great extent for a wireless network, while NOMA gives access to more users with resource constraints. NS provides
users with a better quality of service and network operators with more flexibility. In this work, we propose an NS technique enabled with
NOMA for a MEC network. The proposed NS technique improves service latency for MEC users and reduces the unnecessary allocation
of radio resources in NOMA. The saved resources can be leveraged to accommodate more users, thus increasing the spectral efficiency
of the network. We consider different types of services based on the task completion time of users in this work. The primary focus is to
optimize the total energy consumption for wireless uplink transmission for the NOMA-enabled sliced MEC network. We also propose a
heuristic algorithm as an alternative to reduce the time and computational complexities of the optimization algorithm and simulate the
results extensively to show the effectiveness of our proposed algorithm.
Index Terms—Mobile Edge Computing, Network Slicing, NOMA, Uplink Energy Consumption Minimization, Energy Efficiency, Spectral
Efficiency.
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
2 IEEE TRANSACTIONS ON CLOUD COMPUTING
in this work ensures better management of resources to al. [14] focused on optimizing the offloading, resource al-
improve QoS and reduce the redundant allocations of radio location, and content caching problems for a MEC network
resources in NOMA with various services. to maximize the total revenue. Besides, Bi and Zhang [15]
proposed a MEC model equipped with energy harvesting to
1.2 Contributions provision either complete local computing or full offloading
of a task in order to maximize the aggregate computation
Several works have considered the joint allocation of com- rate. None of the above mentioned works has considered
puting and communication resources with NOMA as the NOMA as an access technique to serve users with radio
access technique for uplink communication for MEC. NS has resources.
also become one of the cutting edge topics for research. Nu- Kiani and Ansari [16] proposed a NOMA-aware MEC
merous works have been published on NS. However, these technique by jointly allocating computing and communica-
works have not included the idea of leveraging NS with tion resources in which users are given access to the radio
NOMA and MEC. In this paper, we propose an NS tech- resource via NOMA. They studied the complete offloading
nique by jointly allocating computing and communication of a task to the MEC server. The main goal of the work is to
resources for uplink in a NOMA-enabled MEC network. optimize the energy consumption for wireless transmission.
Note that we do not consider the downlink transmission Song et al. [17] also proposed an energy-efficient MEC het-
for MEC in this work as the outputs of the tasks that are erogeneous network using NOMA and considered both the
computed in the MEC server are relatively small compared partial and complete offloading of a task to the server. They
to the size of the task to be computed, and the downlink aimed to optimize the total energy consumption, i.e., energy
transmission time is much lower than uplink transmission required for both computing and communication. Pan et al.
time. Major contributions of the work are noted below. [18] proposed NOMA-based uplink and downlink transmis-
1) We propose a novel NS technique for a NOMA-enabled sion for an energy-efficient MEC network. They optimized
MEC network to optimize the energy consumption of transmit powers, transmission time allocation, and task
the uplink communication of the system. We aim to offloading partitions. Yang et al. [19] proposed an energy-
increase both energy efficiency and spectral efficiency efficient NOMA-based MEC network to minimize the com-
of the MEC network by deploying NOMA for uplink bination of completion time and energy consumption of
communication in the MEC network. uplink communication. Nouri et al. [20] proposed another
2) We introduce a novel idea to reduce the redundant optimization mechanism for energy consumption and task
allocation of resources by including NS for NOMA completion delay in an ultradense Internet of Things (UD-
clustering. We also consider NS for a virtual MEC IoT) network by leveraging NOMA. Furthermore, Huang et
network where the virtual computing resources are al. [21] introduced a NOMA-aided MEC network to offload
assigned to the users according to the requirements of a task both directly to the server and with the help of
the respective slices. another user simultaneously. The main objective of the work
3) We propose an optimization problem by incorporating is to minimize energy consumption and maximize data
the slicing technique for user clustering for different offloading. The above works have demonstrated the vitality
applications, joint allocation of computing and commu- of NOMA and MEC, but they did not considered NS.
nication resources, and transmission power control of A number of conceptual architectures by leveraging NS
users for the MEC network. The slices are created based for different network scenarios, including virtual cloud and
on MEC users’ task completion times. edge computing, have been proposed recently [22], [23].
4) To reduce the time and computational complexities Zhang et al. [24] discussed the challenges and opportunities
of the proposed optimization problem, we propose of a sliced logical network to manage the resources and
a heuristic algorithm, referred to as Slicing Aware mobility of users. Likewise, Sharma et al. [25] discussed the
Clustering and Resource Allocation (SACRA) algo- NS method for virtual computing both in cloud and edge
rithm, which is partitioned into three separate algo- of a 5G network by leveraging a flexible heterogeneous
rithms for user clustering, computing resource alloca- network architecture. These works include the idea of NS
tion, and wireless resource allocation. and computing but do not employ the concept of NS with
5) We verify the effectiveness of SACRA by extensive NOMA and MEC.
simulations and comparison with optimal solutions. We organize the rest of the paper as follows. Section 2
shows the system model of our proposed work. We mainly
discuss the categorizing process of users into slices and
1.3 Related Works
NOMA clusters as well as the resource allocation. We for-
MEC is an enabling technology for 5G and beyond net- mulate the optimization problem in Section 3, while Section
works, and NOMA has recently been proposed to improve 4 describes our proposed SACRA algorithm. We present
spectral efficiency. Shi et al. [11] discussed the main chal- the extensive numerical results for our proposed work in
lenges and opportunities of edge computing for future gen- Section 5, followed by the conclusion in Section 6.
eration wireless networks. The work provides an excellent
overview of the MEC network while Ceselli et al. [12] pro-
posed a model to optimize the holistic network architecture 2 S YSTEM M ODEL
for MEC. Chen et al. [13] proposed a distributed mechanism Our system model mainly comprises the NS technique to
to offload the computation task efficiently to the MEC server categorize users for different types of services, NOMA for
by considering both local and edge computing. Wang et giving access to users, and edge computing to reduce the
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
HOSSAIN AND ANSARI: NETWORK SLICING FOR NOMA ENABLED EDGE COMPUTING 3
latency time of the required services of users in a network. to cluster ck as well as in slice sj and otherwise = 0. υij,k
Now, we assume that our MEC network is empowered by Here, users are partitioned into a total |S| slices and |C|
both SDN and NFV functionalities to support NS. It is well- NOMA clusters. Following the message decoding method
known that SDN separates the data plane from the control of NOMA, successive interference cancellation (SIC), users
plane to make a network programmable and NFV helps need to be arranged in an order in a NOMA cluster [6].
to virtualize the physical resources into logical resources Usually, users are organized in descending order for uplink
of a system [1]. However, we consider a single macrocell NOMA inside a NOMA cluster based on their channel
with a base station (BS), where the BS contains a MEC gains. In this work, we denote a user with a higher channel
server to facilitate computing and user devices are equipped gain and that with a lower channel gain as a stronger
with single antennae. The neighboring cells of the macrocell user and a weaker user, respectively. In a NOMA cluster, a
use different frequency bands for communication, and thus stronger user and a weaker user are assigned to lower order
nullifying intercell interference in our proposed model. We and higher order, respectively. In other words, the strongest
also assume that users impose different latency demands for user in a NOMA cluster is given the first position in the
various computing tasks. At this point, we categorize users order; therefore, messages of the strongest user is decoded
for different slices based on the latency demands to deploy first by SIC.
the NS method in our proposed MEC model. Denote m as the order of a user in a cluster and M
We assume that the MEC server can serve all the users as the maximum number of users permitted in a NOMA
situated in the cell. The physical computing resources of cluster. So, there are a total of M orders of users in a
the server are transformed into virtual computing resources NOMA cluster. Theoretically, M can be any number greater
as logical processors, i.e., virtual CPUs (vCPUs). vCPUs are than 1 for NOMA since at least two users use the same
allocated to the users by creating virtual machines (VMs) frequency RBs. Still, M depends on the capability of SIC,
with the help of a hypervisor. We also assume that the which becomes complex and delay-prone with the increase
memory size of the MEC server is sufficient to accomplish in M . Let MT h be the maximum number of users in a
all the tasks. Our proposed model facilitates access to ra- NOMA cluster such that SIC can decode their messages
dio resources by means of NOMA while computing and successfully, and therefore, MT h ≥ 2 and M ≤ MT h .
wireless resources are allocated based on the requirements Denote user ui being assigned to the m-th order of NOMA
k,m
of slices. We divide our remaining system model into two cluster ck by the binary variable ϕi = 1, and 0 otherwise.
subsections: (a) user grouping for slice and NOMA cluster As mentioned earlier, SIC in uplink NOMA decodes the
and (b) resource allocation. We also tabulate the significant messages of a stronger user before the messages of the
notations used throughout the article in Table 1. weaker users in a NOMA cluster. Consequently, stronger
users get interference from weaker users. In comparison,
2.1 User Grouping for Slice and NOMA Cluster weaker users do not get any interference from strong users.
Denote U = {ui |i = 1, ...., |U|} as the set of users in the 2.2 Resource Allocation
macro-cell. Each user is not allowed to offload more than We consider allocation of two types of resources: (i) comput-
one task to the MEC server until the task is completed. We ing resource allocation and (ii) wireless resource allocation.
denote the required latency, required CPU cycle, and the
data size of the task of user ui as Li , Fi , and Di , respectively. 2.2.1 Computing Resources Allocation
We denote the set of slices as S = {sj |j = 1, ...., |S|} where We propose to first allocate the users with computing re-
each slice has a lower and an upper bound of latency to sources according to their order indexes in the NOMA
categorize the users based on their required latencies. We clusters. The allocation process ensures the separation of
also assume that the lower and upper latency bounds for computing resources for slices as NOMA clusters are not
every slice are different. Now, we denote the minimum and shared between users of different slices. The physical com-
maximum latency bound of slice sj as Lmin j and Lmax
j , re- puting resources are assigned to users as virtual resources,
spectively. The maximum latency bound of a slice becomes i.e., vCPUs by creating VMs in the server. VMs are created
the minimum latency bound of the next slice if we arrange with the help of a hypervisor, a software, sometimes called
the slices in ascending order based on their maximum virtual machine monitor (VMM). Both the creation of VMs
latency values. Here, we set the minimum latency as 0 and the allocation of vCPUs depend on the properties of
for the slice with the lowest maximum latency bound, i.e., physical CPUs too. It is also worth noting that each thread
Lmin
j0 = 0, where j 0 = argminsj ∈S [Lmax j ]. Therefore, user ui of each core of a CPU can serve as a vCPU while the total
with Lmin
j < Li ≤ L max
j is assigned to slice sj . number of vCPUs in a server can be counted by multiplying
As we use NOMA for wireless resource allocation, more of the number of CPU sockets, physical cores per socket, and
than one user can use the same frequency resource and the the number of threads per core in the server.
users who use the same frequency resource form a NOMA To simplify the allocation process, we assume one vCPU
cluster. The granularity of the frequency resource that is as the granularity of computing resources to execute the task
used by users of a NOMA cluster is called a frequency of a user. We consider vCPUs to be homogeneous, i.e., the
resource block (RB). Denote C = {ck |k = 1, ...., |C|} as clock speeds of vCPUs are the same. Now, we assume the
the set of NOMA clusters considering all the slices of the clock speed of a vCPU is X cycle per second. Without loss
system. In our proposed model, we consider that users from of generality, a VM is created on-demand for each user by
j,k
different slices do not share the same clusters. Denote υi the VMM and vCPUs are allocated to the VM to execute a
j,k
as a binary variable with υi = 1 when user ui is assigned task. The VM uses multi-threading to execute the task of a
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
4 IEEE TRANSACTIONS ON CLOUD COMPUTING
user. In other words, the task is shared among the vCPUs TABLE 1
of a VM. Denote the set of computing resources (vCPUs) as Significant Notations
V = {vl |l = 1, ...., |V|}. However, we assign the computing Notation Description
resources to the users by considering both NOMA clusters hn
i Channel gain of user ui over frequency RB bn
and order indexes in NOMA clusters. For this purpose, we pn
i Transmission power of user ui through
use the binary variable ψ l,k,m to assign vCPU vl to the m-th frequency RB bn
order of the ck -th NOMA cluster. Therefore, the computing ϕik,m Decision variable to assign user ui to the m-th
time, Qi , for the task of user ui can be expressed as follows. order of NOMA cluster ck
χn,k Decision variable to allocate frequency RB bn
M to NOMA cluster ck
Fi
ϕk,m
X X X
Qi = i l,k,m
(1) υij,k Decision variable to assign user ui to NOMA
ck ∈C m=1 vl ∈V
ψ X cluster ck and slice sj
ψ l,k,m Decision variable to assign vCPU vl to the
2.2.2 Communication Resources Allocation m-th order of NOMA cluster ck
M Total number of users in a NOMA cluster
Our proposed model allocates the communication resources
W Bandwidth of a frequency RB
after allocating computing resources. We denote the set of X Clock speed of vCPU
available frequency RBs, B = {bn |n = 1, ...., |B|} where Di Data size of the task of user ui
the bandwidth of each RB is W . Let us assume the binary Ei Uplink transmission energy consumption
variable χn,k to indicate the allocation of frequency RB of user ui
bn to NOMA cluster ck . Here, χn,k ensures the partition Fi CPU cycle required for the task of user ui
of frequency RBs for the NOMA clusters as well as for Li Latency-demand of user ui
the slices since users from different slices are not clustered Qi Computing time for the task of user ui
together. At this point, we assume channel condition varies Ri Uplink transmission data rate of user ui
Ti Uplink transmission time of user i
across the frequency RBs and use the variable hn i to indicate MT h Maximum number of users allowed in a
the channel gain for user ui over frequency RB bn . We also NOMA cluster
denote pn i and Pi
T ot
as the transmission power through Lmin Lower bound latency demand of slice sj
j
frequency RB bn and the total power budget of user ui , Lmax Upper bound latency demand of slice sj
j
respectively. Accordingly, the data rate, Ri of user ui , can PiT ot Total power budget of user ui
be expressed as Rimin Minimum data rate required by user ui
M B; |B| Set of frequency RBs; Total number of
ϕk,m frequency RBs
X X X
Ri = i χn,k W
C ; |C| Set of NOMA clusters; Total number of
ck ∈C m=1 bn ∈B
! (2) NOMA clusters
pnq hnq S ; |S| Set of slices; Total number of slices
× log2 1 + PM k,r n n U ; |U | Set of users; Total number of users
σ2
P
+ uq ∈U \i r=m+1 ϕq pq hq
V ; |V| Set of vCPUs; Total number of vCPUs
M Jj Fairness index for computing resource
ϕk,m χn,k indicate user ui can be
P P P
where i and allocation to slice sj
ck ∈C m=1 bn ∈B U j ; |U j | Set of users assigned to slice sj ; Total number
assigned to only one order (m-th order) of a particular of users assigned to slice sj
NOMA cluster ck and a frequency RB bn ∈ B can be al-
located to only one NOMA cluster ck , respectively. Besides, of user. Undoubtedly, Qi is also less than Li when (4) is
σ 2 symbolizes the noise power. The energy consumption, satisfied. Therefore, the achievable data rate of user ui must
Ei , for the uplink transmission for the task of user ui can be be greater than or equal to the required minimum data rate
written as X for user ui , Rimin , as shown in (5) if Ti satisfies (4) too.
Ei = Ti pni (3)
bn ∈B Ti + Qi ≤ Li (4)
where Ti = (Di /Ri ) represents the uplink transmission
time.
Ri ≥ Rimin (5)
3 O PTIMIZATION P ROBLEM
Di
Our proposed optimization problem considers the total Rimin = (6)
Li − Qi
energy consumption of all the users for the uplink com-
munication time. The optimization problem needs to satisfy
the conditions shown in (4) with several other constraints The objective of our optimization problem and the con-
related to grouping users into slices, clustering users for straints of the problem can be formulated as follows.
NOMA clusters, assigning users in the orders of NOMA
clusters, and the allocation of computing and communi- minimize X
P1 : Ei
cation resources. It is evident from (4) that the total time υij,k , ϕk,m
i , ψ l,k,m , χn,k , pni
ui ∈U
for computing and transmission must be less than or equal
to the total latency bound requirement of the application
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
HOSSAIN AND ANSARI: NETWORK SLICING FOR NOMA ENABLED EDGE COMPUTING 5
υij,k
X X
s.t. C1 : = 1 ∀ui ∈ U achievable data rate of a user is greater or equal to the
sj ∈S ck ∈C required data rate of the user. Fundamentally, C11 and C12
M together impose the summation of uplink transmission time
ϕk,m
X X
C2 : i = 1 ∀ui ∈ U and computing time of a task to be kept within the latency
ck ∈C m=1 time bound of the task. However, C13 allocates positive
M transmit power for a frequency RB while C14, C15, C16, and
ϕk,m
X X
jk k,m
C3 : i ≥ 2 ∀ck ∈ C C17 keep υi , , ϕi , ψ l,k,m , and χn,k as binary variables,
ui ∈U m=1 respectively.
M Our optimization problem P1 is a mixed integer non-
ϕk,m
X X
C4 : i ≤ MT h ∀ck ∈ C linear programming (MINLP) problem. P1 is not a convex
ui ∈U m=1 problem; therefore, finding an optimal solution is very dif-
C5 : ϕk,m
i ≤ ϕk,m−1
i ∀ui ∈ U; ck ∈ C; 2 ≤ m ≤ M ficult. Besides, the exhaustive search for finding the optimal
X X M solution incurs vast computational and time complexities,
l,k,m thus making the problem intractable. Therefore, we propose
C6 : ψ ≤ 1 ∀vl ∈ V
ck ∈C m=1 the SACRA algorithm to transform the problem into a
M
X X X tractable one to obtain a near-optimal solution. At first, we
C7 : ψ l,k,m ≤ |V| aim to group and assign the users into slices and to form
vl ∈V ck ∈C m=1 NOMA clusters inside the slices in the algorithm. Then, we
allocate the computing and frequency resources to the users
X
C8 : χn,k ≤ 1 ∀bn ∈ B
ck ∈C
of the NOMA clusters of the slices. The grouping of users
X X and the allocation of resources remove some constraints and
C9 : χn,k ≤ |B| binary variables from our optimization problem.
bn ∈B ck ∈C
X Our resource allocation procedures are able to find out a
C10 : pni ≤ PiT ot ∀ui ∈ U near-optimal transmission time for every user and the near-
bn ∈B optimal transmission time, ultimately, producing a near-
C11 : Qi ≤ Li ∀ui ∈ U optimal solution for the energy consumption of our opti-
C12 : Ri ≥ Rimin ∀ui ∈ U mization problem. At this stage, we control the transmission
n power for the users per NOMA cluster to optimize the
C13 : pi ≥ 0 ∀ui ∈ U; bn ∈ B total energy consumption for the uplink transmission of the
jk
C14 : υi , ∈ {0, 1} ∀ui ∈ U; sj ∈ S; ck ∈ C system. The data rate, which is related to the transmission
power of a user in a NOMA cluster, depends not only on
C15 : ϕk,m
i ∈ {0, 1} ∀ui ∈ U; ck ∈ C; m = 1, .., M
the self transmission power but also on the transmission
C16 : ψ l,k,m ∈ {0, 1} ∀vl ∈ V; ck ∈ C; m = 1, .., M power of other users of the NOMA cluster. Consequently,
C17 : χn,k ∈ {0, 1} ∀bn ∈ B; ck ∈ C the control of transmission power for the users per NOMA
cluster can make the optimization problem P1 into a convex
problem [16]. Thus, the partitioning of the problem will
Here, constraint C1 represents the assignment of a user provide an sub-optimal solution.
to only one slice as well as to one NOMA cluster. C1 also
ensures that the users from different slices are not assigned Lemma 1. Optimization problem P1 can be transformed to
jk k,m
to the same NOMA cluster. C2 assigns a user to one of the problem P2 by relaxing binary variables υi , , ϕi , ψ l,k,m , and
n,k
orders in a NOMA cluster while C3 imposes that it takes χ .
at least two users from a slice to form a NOMA cluster. Proof. The binary variables are relaxed by assigning the
C4 limits the assignment of users in a NOMA cluster up users into slices and NOMA clusters, and by allocating
to the SIC capacity whereas C5 represents the assignment the resources. Now, we control the transmission power
priority of a user to a lower order in a NOMA cluster as individually for each NOMA cluster to get the overall op-
compared to a higher order. To this point, C6 enforces a timized transmission power, which gives us the minimized
vCPU be allocated to a user according to the order of a transmission energy for the system eventually. We consider
NOMA cluster. The sequential allocation of vCPUs helps a particular NOMA cluster to optimize the transmission
obtain the minimum data rate requirement for all the users power of its users. We denote the set of the assigned users
in a NOMA cluster so that the wireless resources can be as Ũ with ui to denote an user of Ũ . Besides, we consider
allocated effectively to a NOMA cluster because all users the set of allocated frequency RBs in our interested NOMA
in a NOMA cluster share the same frequency resources to cluster as B̃ where bn denotes a frequency RB of B̃ . So, we
achieve the required data rates. Besides, C7 maintains the can construct the optimization problem P2 from P1.
limit to assign the computing resources to users. C8 allows
the allocation of a frequency RB to only one NOMA cluster, minimize X X n
P2 : pi
and C9 keeps the total number of assigned frequency RBs pni
ui ∈Ũ bn ∈B̃
to users within the capacity. We use constraint C10 to keep
the total allocated transmission power over all the allocated
frequency RBs within the total budget of the transmission s.t. C1 : Pin ≥ 0 ∀ui ∈ Ũ; bn ∈ B̃
X
power for a user. We introduce C11 to uphold the computing C2 : pni ≤PiT ot ∀ui ∈ Ũ
time within the latency bound while C12 ensures that the bn ∈B̃
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
6 IEEE TRANSACTIONS ON CLOUD COMPUTING
bn ∈B̃ 6: if Lmin
j < Li ≤ Lmaxj then
0 0
exp(Z n n
)h 7: U ← U \ui and Û ← Û ∪ ui
i i
C3 : Yin ≤ W log2 1 + P|Ũ |
8: end if
σ 2 + q=i+1 exp(Zqn )hnq 9: end for
∀ui ∈ Ũ 10: Sort all |Û| users such that h̄1 ≥ h̄2 ≥ .... ≥ h̄|Û |
11: if |Û| ≤ M then
X
C4 : Ri = Yin ∀ui ∈ Ũ 0
12: K ← 1; k ← (K + K);
bn ∈B̃
13: for m = 1 to M do
We add a new constraint C3 in problem P3 by con- 14: Assign m-th user of Û in the m-th order
sidering (7) to show the inequality relation between both of the k -th NOMA cluster of C
sides of C3 in order to maintain the convexity for P3. 15: end for
Here, C3 does not invalidate the relation shown in (7) 16: else if (|Û| mod M ) == 0 then
when the solution
is optimal in problem
P3. The data rate 17: K ← (|Û|/M )
exp(Zin )hn 18: for k̂ = 1 to K do
from W log2 1 + P|Ũ |
i
, which is basically 0
σ 2 + q=i+1 exp(Zqn )hn
q 19: k ← (K + k̂);
the Shannon capacity formula, cannot be less than Yin but
20: for m = 1 to M do
equality is achieved when the solution is optimal.
21: Assign {k̂ + (m − 1)K}-th user of Û to the
Lemma 3. The objective function, C1, and C2 of problem P3 are m-th order of the k -th NOMA cluster of C
convex but C3 and C4 are not convex. C3 and C4 become convex 22: end for
when the SINR is high enough such that log(1 + SIN R) ≈ 23: end for
log(SIN R). 24: else
25: if M < MjT h thenk
Proof. The objective function of P3 is a sum of the expo-
26: K ← |Û|/M
nential functions, exp(Zin ), and is thus convex. Besides,
constraints C1 and C2 show the convexity because of the 27: M ← (M + 1)
exponential function or sum of the exponential functions. 28: Do same as in lines 18-23
Still, C3 is not convex because of log(1 + s) where s de- 29: else l m
notes the SINR of throughput. However, we can assume 30: K ← |Û|/M
log(1+s) ≈ log(s) when the SINR becomes high. Therefore, 31: Do same as in lines 18-23
C3 can be transformed as follows. 32: end if
−1
|P
Ũ | 33: end if
C3 : Yin ≤ −W log2 {σ 2 exp(−Zin )hn
i + exp(Zqn 34: K 0 ← (K 0 + K)
q=i+1
−1 35: end for
−Zin )hni hnq } ∀ui ∈ Ũ
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
HOSSAIN AND ANSARI: NETWORK SLICING FOR NOMA ENABLED EDGE COMPUTING 7
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
8 IEEE TRANSACTIONS ON CLOUD COMPUTING
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
HOSSAIN AND ANSARI: NETWORK SLICING FOR NOMA ENABLED EDGE COMPUTING 9
50 RBs
Energy Efficiency (Mbit/Joule)
SACRA Algorithm 25
25 RBs
80
20
60
15
40
10
20 5
0 0
12 18 24 30 36 12 18 24 30 36
Number of Users Number of Users
Fig. 1. Comparison of wireless energy consumption between the opti- Fig. 3. Comparison of EE by varying the number of users over different
mization algorithm and the SACRA algorithm. numbers of frequency RBs.
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
10 IEEE TRANSACTIONS ON CLOUD COMPUTING
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCC.2021.3117754, IEEE
Transactions on Cloud Computing
HOSSAIN AND ANSARI: NETWORK SLICING FOR NOMA ENABLED EDGE COMPUTING 11
[9] C. Marquez et al., “Resource sharing efficiency in network slicing,” Mohammad Arif Hossain (S’19) received the
IEEE Transactions on Network and Service Management, vol. 16, no. 3, B.Sc. degree in electrical and electronic engi-
pp. 909–923, 2019. neering from Khulna University of Engineering
[10] Q. Liu, T. Han, and N. Ansari, “Learning-assisted secure end-to- and Technology, Khulna, Bangladesh and the
end network slicing for cyber-physical systems,” IEEE Network, M.Sc. degree in electronics engineering from
vol. 34, no. 3, pp. 37–43, 2020. Kookmin University, Seoul, South Korea. He is
[11] W. Shi et al., “Edge computing: Vision and challenges,” IEEE currently pursuing the Ph.D. degree in computer
Internet of Things Journal, vol. 3, no. 5, pp. 637–646, 2016. engineering at the New Jersey Institute of Tech-
[12] A. Ceselli, M. Premoli, and S. Secci, “Mobile edge cloud net- nology, Newark, NJ, USA. His current research
work design optimization,” IEEE/ACM Transactions on Networking, interests include cloud and edge computing,
vol. 25, no. 3, pp. 1818–1831, 2017. NOMA, network slicing, network optimization,
[13] X. Chen et al., “Efficient multi-user computation offloading for and machine learning.
mobile-edge cloud computing,” IEEE/ACM Transactions on Net-
working, vol. 24, no. 5, pp. 2795–2808, 2016.
[14] C. Wang et al., “Computation offloading and resource allocation
in wireless cellular networks with mobile edge computing,” IEEE
Transactions on Wireless Communications, vol. 16, no. 8, pp. 4924–
4938, 2017.
[15] S. Bi and Y. J. Zhang, “Computation rate maximization for wire-
less powered mobile-edge computing with binary computation
offloading,” IEEE Transactions on Wireless Communications, vol. 17,
no. 6, pp. 4177–4190, 2018.
[16] A. Kiani and N. Ansari, “Edge computing aware NOMA for 5G
networks,” IEEE Internet of Things Journal, vol. 5, no. 2, pp. 1299–
1306, 2018.
[17] Z. Song, Y. Liu, and X. Sun, “Joint radio and computational
resource allocation for NOMA-based mobile edge computing in
heterogeneous networks,” IEEE Communications Letters, vol. 22,
no. 12, pp. 2559–2562, 2018.
[18] Y. Pan, et al., “Energy-efficient NOMA-based mobile edge com-
Nirwan Ansari (S’78–M’83–SM’94–F’09), Dis-
puting offloading,” IEEE Communications Letters, vol. 23, no. 2, pp.
tinguished Professor of Electrical and Computer
310–313, 2019.
Engineering at the New Jersey Institute of Tech-
[19] Z. Yang, C. Pan, J. Hou, and M. Shikh-Bahaei, “Efficient resource
nology (NJIT), received his Ph.D. from Purdue
allocation for mobile-edge computing networks with NOMA:
University, MSEE from the University of Michi-
Completion time and energy minimization,” IEEE Transactions on
gan, and BSEE (summa cum laude with a per-
Communications, vol. 67, no. 11, pp. 7771–7784, 2019.
fect GPA) from NJIT. He is also a Fellow of
[20] N. Nouri et al., “Joint access and resource allocation in ultradense
National Academy of Inventors.
mmWave NOMA networks with mobile edge computing,” IEEE
He authored Green Mobile Networks: A Net-
Internet of Things Journal, vol. 7, no. 2, pp. 1531–1547, 2020.
working Perspective (Wiley-IEEE, 2017) with T.
[21] Y. Huang, Y. Liu, and F. Chen, “NOMA-aided mobile edge com-
Han, and co-authored two other books. He has
puting via user cooperation,” IEEE Transactions on Communications,
also (co-)authored more than 600 technical publications. He has guest-
vol. 68, no. 4, pp. 2221–2235, 2020.
edited a number of special issues covering various emerging top-
[22] A. Ksentini and P. A. Frangoudis, “Toward slicing-enabled multi-
ics in communications and networking. He has served on the edito-
access edge computing in 5g,” IEEE Network, vol. 34, no. 2, pp.
rial/advisory board of over ten journals including as Associate Editor-
99–105, 2020.
in-Chief of IEEE Wireless Communications Magazine. His current re-
[23] Z. Mlika and S. Cherkaoui, “Network slicing with mec and deep
search focuses on green communications and networking, cloud com-
reinforcement learning for the internet of vehicles,” IEEE Network,
puting, drone-assisted networking, and various aspects of broadband
pp. 1–7, 2021.
networks.
[24] H. Zhang et al., “Network slicing based 5G and future mobile
He was elected to serve in the IEEE Communications Society (Com-
networks: Mobility, resource management, and challenges,” IEEE
Soc) Board of Governors as a member-at-large, has chaired some Com-
Communications Magazine, vol. 55, no. 8, pp. 138–145, 2017.
Soc technical and steering committees, is current Director of ComSoc
[25] S. Sharma, R. Miller, and A. Francini, “A cloud-native approach to
Educational Services Board, has been serving in many committees
5G network slicing,” IEEE Communications Magazine, vol. 55, no. 8,
such as the IEEE Fellow Committee, and has been actively organizing
pp. 120–127, 2017.
numerous IEEE International Conferences/Symposia/Workshops. He is
[26] D. Julian, M. Chiang, D. O’Neill, and S. Boyd, “QoS and fair-
frequently invited to deliver keynote addresses, distinguished lectures,
ness constrained convex optimization of resource allocation for
tutorials, and invited talks. Some of his recognitions include several
wireless cellular and ad hoc networks,” in Proceedings.Twenty-First
excellence in teaching awards, a few best paper awards, the NCE Ex-
Annual Joint Conference of the IEEE Computer and Communications
cellence in Research Award, several ComSoc TC technical recognition
Societies, vol. 2, 2002, pp. 477–486.
awards, the NJ Inventors Hall of Fame Inventor of the Year Award,
[27] Spatial channel model for Multiple Input Multiple Output (MIMO)
the Thomas Alva Edison Patent Award, Purdue University Outstanding
simulations, 3GPP TR 25.996 version 15.0.0 Release 15, 3GPP, 2018,.
Electrical and Computer Engineering Award, the NCE 100 Medal, NJIT
[28] LTE; Evolved Universal Terrestrial Radio Access (E-UTRA); User
Excellence in Research Prize and Medal, and designation as a COM-
Equipment (UE) radio transmission and reception, 3GPP TS 36.101
SOC Distinguished Lecturer. He has also been granted more than 40
version 14.7.0 Release 14, 3GPP, 2018,.
U.S. patents.
2168-7161 (c) 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Istanbul Universitesi. Downloaded on March 31,2022 at 11:56:55 UTC from IEEE Xplore. Restrictions apply.