Professional Documents
Culture Documents
Energy Efficient Federated Learning Over Wireless Communication Networks
Energy Efficient Federated Learning Over Wireless Communication Networks
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
1936 IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL. 20, NO. 3, MARCH 2021
II. S YSTEM M ODEL where κ is the effective switched capacitance that depends on
the chip architecture. To compute the local FL model, user k
Consider a cellular network that consists of one BS serving needs to compute Ck Dk CPU cycles with Ik local iterations,
K users, as shown in Fig. 1. Each user k has a local dataset Dk which means that the total computation energy at user k is
with Dk data samples. For each dataset Dk = {xkl , ykl }D l=1 ,
k
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
YANG et al.: ENERGY EFFICIENT FEDERATED LEARNING OVER WIRELESS COMMUNICATION NETWORKS 1937
transmit data of size s within a time duration tk , the wireless min F (w) Fk (w) = f (w, xkl , ykl ),
w D D
transmit energy of user k will be: EkT = tk pk . k=1 k=1 l=1
(9)
3) Information Broadcast: In this step, the BS aggregates
K
the global FL model. The BS broadcasts the global FL model where D = k=1 Dk is the total data samples of all users.
to all users in the downlink. Due to the high transmit power To solve problem (9), we adopt the distributed approximate
at the BS and the high bandwidth that can be used for data Newton (DANE) algorithm in [25], which is summarized in
broadcasting, the downlink time is neglected compared to the Algorithm 1. Note that the gradient descent (GD) is used at
uplink data transmission time. It can be observed that the local each user in Algorithm 1. One can use stochastic gradient
data Dk is not accessed by the BS, so as to protect the privacy descent (SGD) to decrease the computation complexity for
of users, as is required by FL. cases in which a relatively low accuracy can be tolerated.
According to the above FL model, the energy consumption However, if high accuracy was needed, gradient descent (GD)
of each user includes both local computation energy EkC and would be preferred [25]. Moreover, the number of global
wireless transmission energy EkT . Denote the number of global iterations is higher in SGD than in GD. Due to limited wireless
iterations by I0 , and the total energy consumption of all users resources, SGD may not be efficient since SGD requires
that participate in FL will be more iterations for wireless transmissions compared to GD.
K The DANE algorithm is designed to solve a general local
E = I0 (EkC + EkT ). (5) optimization problem, before averaging the solutions of all
k=1 users. The DANE algorithm relies on the similarity of the
Hereinafter, the total time needed for completing the exe- Hessians of local objectives, representing their iterations as
cution of the FL algorithm is called completion time. The an average of inexact Newton steps. In Algorithm 1, we can
completion time of each user includes the local computation see that, at every FL iteration, each user downloads the
time and transmission time, as shown in Fig. 3. Based on (1), global FL model from the BS for local computation, while
the completion time Tk of user k will be the BS periodically gathers the local FL model from all
users and sends the updated global FL model back to all
Ik Ck Dk users. According to Algorithm 1, since the local optimization
Tk = I0 (τk + tk ) = I0 + tk . (6)
fk problem is solved with several local iterations at each global
Let T be the maximum completion time for training the entire iteration, the number of full gradient computations will be
FL algorithm and we have smaller than the number of local updates.
We define w(n) as the global FL model at a given iter-
Tk ≤ T, ∀k ∈ K. (7) ation n. In practice, each user solves the local optimization
problem
B. FL Model
min Gk (w (n) , hk ) Fk (w (n) + hk )
We define a vector w to capture the parameters related hk ∈Rd
to the global FL model. We introduce the loss function − (∇Fk (w (n) ) − ξ∇F (w (n) ))T hk . (11)
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
1938 IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL. 20, NO. 3, MARCH 2021
(n),(i)
Algorithm 1 FL Algorithm hk of problem (11) with accuracy η means that
1: Initialize global model w 0 and global iteration number n = (n),(i) (n)∗
Gk (w(n) , hk ) − Gk (w(n) , hk )
0. (n),(0) (n)∗
(n)
2: repeat ≤ η(Gk (w , hk ) − Gk (w (n) , hk )), (14)
3: Each user k computes ∇Fk (w (n) ) and sends it to the (n)∗
where hk is the optimal solution of problem (11). Each
BS. K user is assumed to solve the local optimization problem (11)
4: The BS computes ∇F (w (n) ) = 1 K k=1 ∇Fk (w
(n)
),
with a target accuracy η. Next, in Lemma 1, we derive a lower
which is broadcast to all users.
bound on the number of local iterations needed to achieve a
5: parallel for user k ∈ K = {1, · · · , K}
local accuracy η in (14).
6: Initialize the local iteration number i = 0 and set 2
(n),(0) Lemma 1: Let v = (2−Lδ)δγ . If we set step δ < L2 and run
hk = 0.
the gradient method
7: repeat
(n),(i+1) (n),(i) (n),(i)
8: Update hk = hk − δ∇Gk (w(n) , hk ) i ≥ v log2 (1/η) (15)
and i = i + 1.
9: until the accuracy η of local optimization problem (11) iterations at each user, we can solve local optimization prob-
is obtained. lem (11) with an accuracy η.
10:
(n)
Denote hk = hk
(n),(i) (n)
and each user sends hk to Proof: See Appendix A.
the BS. The lower bounded derived in (15) reflects the growing
11: end for trend for the number of local iterations with respect to
12: The BS computes accuracy η. In the following, we use this lower bound to
approximate the number of iterations Ik needed for local
computations by each user.
1 (n)
K
w (n+1) = w (n) + hk , (10) In Algorithm 1, the iterative method involves a number of
K global iterations (i.e., the value of n in Algorithm 1) to achieve
k=1
a global accuracy 0 for the global FL model. In other words,
and broadcasts the value to all users.
the solution w(n) of problem (9) with accuracy 0 is a point
13: Set n = n + 1.
such that
14: until the accuracy 0 of problem (9) is obtained.
F (w (n) ) − F (w∗ ) ≤ 0 (F (w (0) ) − F (w ∗ )), (16)
where w∗ is the actual optimal solution of problem (9).
In problem (11), ξ is a constant value. Any solution hk of To analyze the convergence rate of Algorithm 1, we make
problem (11) represents the difference between the global FL the following two assumptions on the loss function.
model and local FL model for user k, i.e., w(n) + hk is the
- A1: Function Fk (w) is L-Lipschitz, i.e., ∇2 Fk (w)
local FL model of user k at iteration n. The local optimization
LI.
problem in (11) depends only on the local data and the gradient
- A2: Function Fk (w) is γ-strongly convex,
of the global loss function. The local optimization problem is
i.e., ∇2 Fk (w) γI.
then solved, and the updates from individual users are averaged
to form a new iteration. This approach allows any algorithm The values of γ and L are determined by the loss function.
to be used to solve the local optimization problem. Note that These assumptions can be easily satisfied by widely used
the number of iterations needed to upload the local FL model FL loss functions such as linear or logistic loss functions [18].
(i.e., the number of global iterations) in Algorithm 1 is always Under assumptions A1 and A2, we provide the following
smaller than in the conventional FL algorithms that do not theorem about convergence rate of Algorithm 1 (i.e., the num-
consider a local optimization problem (11) [25]. To solve the ber of iterations needed for Algorithm 1 to converge), where
local optimization problem in (11), we use the gradient method each user solves its local optimization problem with a given
accuracy.
(n),(i+1) (n),(i) (n),(i)
hk = hk − δ∇Gk (w(n) , hk ), (12) Theorem 1: If we run Algorithm 1 with 0 < ξ ≤ L γ
for
(n),(i)
a
where δ is the step size, hk is the value of hk n≥ (17)
1−η
at the i-th local iteration with given vector w (n) , and
(n),(i) 2
∇Gk (w (n) , hk ) is the gradient of function Gk (w (n) , hk ) iterations with a = 2L 1
γ 2 ξ ln 0 , we have F (w
(n)
) − F (w∗ ) ≤
(n),(i)
at point hk = hk . For small step size δ, based on equation 0 (F (w (0) ) − F (w ∗ )).
(n),(0) (n),(1)
(12), we can obtain a set of solutions hk , hk , Proof: See Appendix B.
(n),(2) (n),(i) From Theorem 1, we observe that the number of global
hk , · · · , hk , which satisfy,
iterations n increases with the local accuracy η at the rate
(n),(0) (n),(i) of 1/(1 − η). From Theorem 1, we can also see that the FL
Gk (w(n) , hk ) ≥ · · · ≥ Gk (w (n) , hk ). (13)
performance depends on parameters L, γ, ξ, 0 and η. Note
To provide the convergence condition for the gradient method, that the prior work in [36, Eq. (9)] only studied the number of
we introduce the definition of local accuracy, i.e., the solution iterations needed for FL convergence under the special case in
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
YANG et al.: ENERGY EFFICIENT FEDERATED LEARNING OVER WIRELESS COMMUNICATION NETWORKS 1939
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
1940 IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL. 20, NO. 3, MARCH 2021
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
YANG et al.: ENERGY EFFICIENT FEDERATED LEARNING OVER WIRELESS COMMUNICATION NETWORKS 1941
O(Lit SK), where Lit is the number of iterations for iteratively (1 − η)T Ak log2 η
s.t. tk ≤ + , ∀k ∈ K, (36a)
optimizing (t, η) and (T, b, f , p), and S = log 2(1/1 ) + a fkmax
log 2(1/2 ) log 2(1/3 ). s gk pmax
The conventional successive convex approximation (SCA) ≤ bk log2 1 + k
, ∀k ∈ K, (36b)
tk N0 bk
method can be used to solve problem (19). The complexity of
K
SCA method is O(Lsca K 3 ), where Lsca is the total number of bk ≤ B, (36c)
iterations for SCA method. Compared to SCA, the proposed k=1
Algorithm 3 grows linearly with the number of users K. 0 ≤ η ≤ 1, (36d)
It should be noted that Algorithm 3 is done at the BS side
tk ≥ 0, bk ≥ 0, ∀k ∈ K. (36e)
before executing the FL scheme in Algorithm 1. To implement
Algorithm 3, the BS needs to gather the information of gk , Next, we provide the sufficient and necessary condition for
pmax
k , fkmax , Ck , Dk , and s, which can be uploaded by the feasibility of set (36a)-(36e).
all users before the FL process. Due to small data size, Lemma 3: With a fixed T , set (36a)-(36e) is nonempty if
the transmission delay of these information can be neglected. an only if
The BS broadcasts the obtained solution to all users. Since the
K
BS has high computation capacity, the latency of implementing B ≥ min uk (vk (η)), (37)
Algorithm 3 at the BS will not affect the latency of the FL 0≤η≤1
k=1
process.
where
(ln 2)η
IV. A LGORITHM TO F IND A F EASIBLE S OLUTION uk (η) = − (ln 2)N0 η
, (38)
2)N0 η − gk pmax
OF P ROBLEM (19) W − (ln
gk pmax e k + (ln 2)N0 η
gk pmax
k k
Since the iterative algorithm to solve problem (19) requires
an initial feasible solution, we provide an efficient algorithm and
to find a feasible solution of problem (19) in this section. To s
vk (η) = (1−η)T Ak log2 η
. (39)
obtain a feasible solution of (19), we construct the completion a + fkmax
time minimization problem
Proof: See Appendix F.
min T, (35) To effectively solve (37) in Lemma 3, we provide the
T,t,b,f ,p,η
following lemma.
s.t. (19a) − (19g). (35a)
Lemma 4: In (38), uk (vk (η)) is a convex function.
We define (T ∗ , t∗ , b∗ , f ∗ , p∗ , η ∗ ) as the optimal solution of Proof: See Appendix G.
problem (35). Then, (t∗ , b∗ , f ∗ , p∗ , η ∗ ) is a feasible solution Lemma 4 implies that the optimization problem in (37) is a
of problem (19) if T ∗ ≤ T , where T is the maximum delay convex problem, which can be effectively solved. By finding
in constraint (19a). Otherwise, problem (19) is infeasible. the optimal solution of (37), the sufficient and necessary
Although the completion time minimization problem (35) is condition for the feasibility of set (36a)-(36e) can be simplified
still nonconvex due to constraints (19a)-(19b), we show that using the following theorem.
the globally optimal solution can be obtained by using the Theorem 4: Set (36a)-(36e) is nonempty if and only if
bisection method.
K
B≥ uk (vk (η ∗ )), (40)
A. Optimal Resource Allocation k=1
K
Lemma 2: Problem (35) with T < T ∗ does not have a where η ∗ is the unique solution to
k=1 uk (vk (η ))
∗
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
1942 IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL. 20, NO. 3, MARCH 2021
B. Complexity Analysis
The major complexity of Algorithm 4 at each iteration
lies in checking the feasibility condition (40). To check the
inequality in (40), the optimal η ∗ needs to be obtained by
using the bisection method, which involves the complexity
of O(K log2 (1/4 )) with accuracy 4 . As a result, the total
complexity of Algorithm 4 is O(K log2 (1/4 ) log2 (1/5 )), Fig. 5. Comparison of exact number of required iterations with the upper
where 5 is the accuracy of the bisection method used in bound derived in Theorem 1.
the outer layer. The complexity of Algorithm 4 is low since
O(K log2 (1/4 ) log2 (1/5 )) grows linearly with the total
number of users. user has Dk = 500 data samples, which are randomly selected
Similar to Algorithm 3 in Section III, Algorithm 4 is done from the dataset with equal probability. All statistical results
at the BS side before executing the FL scheme in Algorithm 1, are averaged over 1000 independent runs.
which will not affect the latency of the FL process.
A. Convergence Behavior
V. N UMERICAL R ESULTS Fig. 4 shows the value of the loss function as the number
For our simulations, we deploy K = 50 users uniformly in of total computations varies with IID and non-IID data. For
a square area of size 500 m × 500 m with the BS located at IID data case, all data samples are first shuffled and then
its center. The path loss model is 128.1 + 37.6 log10 d (d is partitioned into 60, 000/500 = 120 equal parts, and each
in km) and the standard deviation of shadow fading is 8 dB. device is assigned with one particular part. For non-IID data
In addition, the noise power spectral density is N0 = −174 case [28], all data samples are first sorted by digit label and
dBm/Hz. We use the real open blog feedback dataset in [39]. then divided into 240 shards of size 250, and each device is
This dataset with a total number of 60,000 data samples assigned with two shards. Note that the latter is a pathological
originates from blog posts and the dimensional of each data non-IID partition way since most devices only obtain two
sample is 281. The prediction task associated with the data is kinds of digits. From this figure, it is observed that the
the prediction of the number of comments in the upcoming FL algorithm with non-IID data shows similar convergence
24 hours. Parameter Ck is uniformly distributed in [1, 3] × 104 behavior with the FL algorithms with IID data. Besides,
cycles/sample. The effective switched capacitance in local we find that the FL algorithm with multiple local updates
computation is κ = 10−28 [35]. In Algorithm 1, we set ξ = requires more total computations than the FL algorithm with
1/10, δ = 1/10, and 0 = 10−3 . Unless specified otherwise, only one local update.
we choose an equal maximum average transmit power pmax 1 = In Fig. 5, we show the comparison of exact number of
· · · = pmax
K = p max
= 10 dB, an equal maximum computation required iterations with the upper bound derived in Theorem 1.
capacity f1max = · · · = fK max
= f max = 2 GHz, a transmit According to this figure, it is found that the gap of the exact
data size s = 28.1 kbits, and a bandwidth B = 20 MHz. Each number of required iterations with the upper bound derived
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
YANG et al.: ENERGY EFFICIENT FEDERATED LEARNING OVER WIRELESS COMMUNICATION NETWORKS 1943
Fig. 6. Completion time versus maximum average transmit power of each Fig. 7. Comparison of completion time with different batch sizes.
user.
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
1944 IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL. 20, NO. 3, MARCH 2021
2
(n),(i)
2
Moreover, the EXH-FDMA scheme achieves almost the same (n),(i) (2 − Lδ)δ
performance as the proposed FL scheme, which shows that = Gk (w (n) , hk )−
2
2
the proposed approach achieves the optimum solution.
(n) (n),(i)
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
YANG et al.: ENERGY EFFICIENT FEDERATED LEARNING OVER WIRELESS COMMUNICATION NETWORKS 1945
i+1
(2 − Lδ)δγ (n)∗ We are new ready to prove Theorem 1. With the above
≤ 1− (Gk (w (n) , 0) − Gk (w (n) , hk ))
2 inequalities and equalities, we have
(2 − Lδ)δγ
≤ exp −(i + 1) (Gk (w (n) , 0) − Gk (w (n) , F (w(n+1) )
2
2
(n)∗
1 L
(n)
K K
hk )), (A.3) (10),A1
(n) (n) T (n)
≤ F (w )+ ∇F (w ) hk +
h k
K 2K 2
where the last inequality follows from the fact that 1 − k=1 k=1
(n),(i)
x ≤ exp(−x). To ensure that Gk (w (n) , hk 1
K
) − (11) (n) (n)
(n)∗ (n)∗ = F (w(n) )+ Gk (w(n) , hk )+∇Fk (w (n) )hk
Gk (w (n) , hk ) ≤ η(Gk (w(n) , 0) − Gk (w(n) , hk )), Kξ
k=1
we have (15).
K
2
(n) L
(n)
− Fk (w (n) + hk ) +
h
A PPENDIX B 2K 2
k
k=1
P ROOF OF T HEOREM 1 K
A2 1 (n)
Before proving Theorem 1, the following lemma is pro- ≤ F (w(n) ) + Gk (w (n) , hk ) − Fk (w (n) )
Kξ
k=1
vided.
K
2
Lemma 5: Under the assumptions A1 and A2, the following γ (n) 2 L
(n)
conditions hold −
hk
+
h k
. (B.8)
2 2K 2
k=1
1
∇Fk (w (n) + h(n) ) − ∇Fk (w(n) )
2
L According to the triangle inequality and mean inequality,
≤ (∇Fk (w(n) + h(n) ) − ∇Fk (w (n) ))T h(n) we have
1
2
≤
∇Fk (w (n) + h(n) ) − ∇Fk (w (n) )
2 , (B.1)
1 K
1
K
2 1
γ
(n)
(n)
(n)
2
h
≤
hk
≤
hk
.
K k
K K
and k=1 k=1 k=1
(B.9)
∇F (w)
2 ≥ γ(F (w) − F (w ∗ )). (B.2)
Combining (B.8) and (B.9) yields
Proof: According to the Lagrange median theorem, there
always exists a w such that F (w (n+1) )
K
(∇Fk (w(n) + h(n) ) − ∇Fk (w (n) )) = ∇2 Fk (w)h(n) .(B.3) 1 (n)
≤ F (w(n) ) + Gk (w(n) , hk ) − Fk (w (n) )
Kξ
Combining assumptions A1, A2, and (B.3) yields (B.1).
k=1
2
γ
F (w∗ ) ≥ F (w) + ∇F (w)T (w ∗ − w) +
w∗ − w
2 . K
2 (14)
(n) 1 γ − Lξ (n) 2
≤ F (w )−
hk
(B.5) Kξ 2
k=1
As a result, we have (n)∗
+ (1 − η)(Gk (w (n) , 0) − Gk (w (n) , hk ))
∇F (w)
2 =
∇F (w) − ∇F (w∗ )
2 K
(11) (n) 1 γ − Lξ (n) 2
(B.4)
≥ γ
∇F (w) − ∇F (w ∗ )
w − w ∗
= F (w )−
hk
Kξ 2
k=1
≥ γ(∇F (w) − ∇F (w ∗ ))T (w − w∗ ) (n) (n)∗
+ (1 − η)(Fk (w ) − Fk (w (n) + hk )
= γ∇F (w) (w − w )
T ∗
(n)∗
(B.5) + (∇Fk (w(n) ) − ξ∇F (w (n) ))T hk )
≥ γ(F (w) − F (w∗ )), (B.6)
K
(B.7) 1 γ − Lξ (n) 2
which proves (B.2). = F (w(n) ) −
hk
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
1946 IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL. 20, NO. 3, MARCH 2021
2 = 2 3 e tk bk ≥ 0. (D.3)
∂bk g k tk b k
F (w(n+1) ) − F (w ∗ ) ∂p∗
From (D.3), we can see that ∂bkk is an increasing function
(1 − η)γ 2 ξ ∂p∗ ∂p∗
≤ 1− (F (w (n) ) − F (w ∗ )) of bk . Since limbk →0+ ∂bkk = 0, we have ∂bkk < 0 for 0 <
2L2
n+1 bk < ∞, i.e., p∗k in (D.1) is a decreasing function of bk . Thus,
(1 − η)γ 2 ξ maximum transmit power constraint p∗k ≤ pmax is equivalent
≤ 1− (F (w (0) ) − F (w∗ )) k
2L2 to
(1 − η)γ 2 ξ (ln 2)s
≤ exp −(n + 1) (F (w (0) ) − F (w ∗ )), bk ≥ bmin − ,
2L2 k (ln 2)N0 s
(ln 2)N0 s − gk pmax tk (ln 2)N0 s
(B.16) tk W − gk pmax tk e k + gk pmax
k k
where the last inequality follows from the fact that 1 − x ≤ (D.4)
exp(−x). To ensure that F (w(n+1) )−F (w∗ ) ≤ 0 (F (w (0) )−
where bmin is defined in (32).
F (w ∗ )), we have (17). k
Substituting (D.1) into problem (28), we can obtain
N0 tk bk b st
A PPENDIX C
K
P ROOF OF T HEOREM 2 min 2 k k −1 , (D.5)
b gk
According to (20), transmitting with minimal time is always k=1
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
YANG et al.: ENERGY EFFICIENT FEDERATED LEARNING OVER WIRELESS COMMUNICATION NETWORKS 1947
According to (D.3), problem (D.5) is a convex function, Substituting (F.3) into (36b), we can construct the following
which can be effectively solved by using the Karush-Kuhn- problem
Tucker conditions. The Lagrange function of (D.5) is
K
K min
N0 tk bk b st
K bk (F.4)
b,η
L(b, μ) = 2 k k −1 +μ bk −B , k=1
k=1
gk
k=1 gk pmax
s.t. vk (η) ≤ bk log2 1 + k
, ∀k ∈ K, (F.4a)
(D.6) N0 bk
0 ≤ η ≤ 1, (F.4b)
where μ is the Lagrange multiplier associated with constraint
(D.5a). The first order derivative of L(b, μ) with respect to bk bk ≥ 0, ∀k ∈ K, (F.4c)
is where vk (η) is defined in (39). We can observe that set (36a)-
∂L(b, μ) N0 (ln 2)s (ln 2)s (ln 2)s (36e) is nonempty if an only if the optimal objective value of
= e tk bk
−1− e tk bk
+ μ.
∂bk g k tk tk b k (F.4) is less than B. Since the right hand side of (36b) is an
(D.7) increasing function, (36b) should hold with equality for the
optimal solution of problem (F.4). Setting (36b) with equality,
We define bk (μ) as the unique solution to ∂L(b,μ)
∂bk = 0. Given problem (F.4) reduces to (37).
constraint (D.5b), the optimal b∗k can be founded from (30).
Since the objective function (D.5) is a decreasing function of A PPENDIX G
bk , constrain (D.5a) always holds with equality for the optimal P ROOF OF L EMMA 4
solution, which shows that the optimal Lagrange multiplier is
obtained by solving (34). We first prove that vk (η) is a convex function. To show this,
we define
A PPENDIX E s
P ROOF OF L EMMA 2 φ(η) = , 0 ≤ η ≤ 1, (G.1)
η
Assume that problem (35) with T = T̄ < T ∗ is fea- and
sible, and the feasible solution is (T̄ , t̄, b̄, f̄ , p̄, η̄). Then,
(1 − η)T Ak log2 η
the solution (T̄ , t̄, b̄, f̄ , p̄, η̄) is feasible with lower value of the ϕk (η) = + , 0 ≤ η ≤ 1. (G.2)
a fkmax
objective function than solution (T ∗ , t∗ , b∗ , f ∗ , p∗ , η ∗ ), which
contradicts the fact that (T ∗ , t∗ , b∗ , f ∗ , p∗ , η ∗ ) is the optimal According to (39), we have: vk (η) = φ(ϕk (η)). Then,
solution. the second-order derivative of vk (η) is:
For problem (35) with T = T̄ > T ∗ , we can always con-
struct a feasible solution (T̄ , t∗ , b∗ , f ∗ , p∗ , η ∗ ) to problem (35) vk (η) = φ (ϕk (η))(ϕk (η))2 + φ (ϕk (η))ϕk (η). (G.3)
by checking all constraints. According to (G.1) and (G.2), we have
s 2s
A PPENDIX F φ (η) = − 2
≤ 0, φ (η) = 3 ≥ 0, (G.4)
η η
P ROOF OF L EMMA 3 Ak
ϕk (η) = − ≤ 0. (G.5)
To prove this, we first define function (ln 2)fkmax η 2
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
1948 IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, VOL. 20, NO. 3, MARCH 2021
Applying the increasing function uk (η) on both sides of (G.7) [15] S. P. Karimireddy, S. Kale, M. Mohri, S. J. Reddi, S. U. Stich,
yields and A. T. Suresh, “SCAFFOLD: Stochastic controlled averaging
for federated learning,” 2019, arXiv:1910.06378. [Online]. Available:
http://arxiv.org/abs/1910.06378
θη1 + (1 − θ)η2 ≥ uk (θzk (η1 ) + (1 − θ)zk (η2 )). (G.8) [16] H. B. McMahan, E. Moore, D. Ramage, S. Hampson, and
B. A. Y. Arcas, “Communication-efficient learning of deep networks
Denote η̄1 = zk (η1 ) and η̄2 = zk (η2 ), i.e., we have η1 = from decentralized data,” 2016, arXiv:1602.05629. [Online]. Available:
uk (η̄1 ) and η2 = uk (η̄2 ). Thus, (G.8) can be rewritten as http://arxiv.org/abs/1602.05629
[17] V. Smith, C.-K. Chiang, M. Sanjabi, and A. S. Talwalkar, “Federated
θuk (η̄1 ) + (1 − θ)uk (η̄1 ) ≥ uk (θη̄1 + (1 − θ)η̄2 ), (G.9) multi-task learning,” in Proc. Adv. Neural Inf. Process. Syst., 2017,
pp. 4424–4434.
[18] H. H. Yang, Z. Liu, T. Q. S. Quek, and H. V. Poor, “Scheduling policies
which indicates that uk (η) is a convex function. As a result, for federated learning in wireless networks,” IEEE Trans. Commun.,
we have proven that uk (η) is an increasing and convex vol. 68, no. 1, pp. 317–333, Jan. 2020.
function, which shows [19] S. Wang et al., “Adaptive federated learning in resource constrained
edge computing systems,” IEEE J. Sel. Areas Commun., vol. 37, no. 6,
uk (η) ≥ 0, uk (η) ≥ 0. (G.10) pp. 1205–1221, Jun. 2019.
[20] E. Jeong, S. Oh, J. Park, H. Kim, M. Bennis, and S.-L. Kim, “Multi-hop
federated private data augmentation with sample compression,” 2019,
To show the convexity of uk (vk (η)), we have arXiv:1907.06426. [Online]. Available: http://arxiv.org/abs/1907.06426
[21] J.-H. Ahn, O. Simeone, and J. Kang, “Wireless federated distilla-
uk (vk (η)) = uk (vk (η))(vk (η))2 + uk (vk (η))vk (η) ≥ 0, tion for distributed edge learning with heterogeneous data,” 2019,
(G.11) arXiv:1907.02745. [Online]. Available: http://arxiv.org/abs/1907.02745
[22] M. Chen, M. Mozaffari, W. Saad, C. Yin, M. Debbah, and C. S. Hong,
“Caching in the sky: Proactive deployment of cache-enabled unmanned
according to vk (η) ≥ 0 and (G.10). As a result, uk (vk (η)) is aerial vehicles for optimized quality-of-experience,” IEEE J. Sel. Areas
a convex function. Commun., vol. 35, no. 5, pp. 1046–1061, May 2017.
[23] M. Mozaffari, W. Saad, M. Bennis, and M. Debbah, “Mobile
unmanned aerial vehicles (UAVs) for energy-efficient Internet of Things
R EFERENCES communications,” IEEE Trans. Wireless Commun., vol. 16, no. 11,
pp. 7574–7589, Nov. 2017.
[1] Z. Yang, M. Chen, W. Saad, C. S. Hong, and M. Shikh-Bahaei, “Delay [24] Z. Yang et al., “Joint altitude, beamwidth, location, and bandwidth
minimization for federated learning over wireless communication net- optimization for UAV-enabled communications,” IEEE Commun. Lett.,
works,” in Proc. Int. Conf. Mach. Learn. Workshop, Jul. 2020, pp. 1–7. vol. 22, no. 8, pp. 1716–1719, Aug. 2018.
[2] S. Wang et al., “When edge meets learning: Adaptive control for [25] J. Konečný, H. B. McMahan, D. Ramage, and P. Richtárik,
resource-constrained distributed machine learning,” in Proc. IEEE Conf. “Federated optimization: Distributed machine learning for on-
Comput. Commun. (INFOCOM), Honolulu, HI, USA, Apr. 2018, device intelligence,” 2016, arXiv:1610.02527. [Online]. Available:
pp. 63–71. http://arxiv.org/abs/1610.02527
[3] Y. Sun, M. Peng, Y. Zhou, Y. Huang, and S. Mao, “Application of [26] M. M. Amiri and D. Gunduz, “Machine learning at the wireless edge:
machine learning in wireless networks: Key techniques and open issues,” Distributed stochastic gradient descent over-the-air,” in Proc. IEEE Int.
IEEE Commun. Surveys Tuts., vol. 21, no. 4, pp. 3072–3108, Jun. 2019. Symp. Inf. Theory (ISIT), Paris, France, Jul. 2019, pp. 1432–1436.
[4] Y. Liu, S. Bi, Z. Shi, and L. Hanzo, “When machine learning meets big [27] G. Zhu, D. Liu, Y. Du, C. You, J. Zhang, and K. Huang,
data: A wireless communication perspective,” 2019, arXiv:1901.08329. “Towards an intelligent edge: Wireless communication meets
[Online]. Available: http://arxiv.org/abs/1901.08329 machine learning,” 2018, arXiv:1809.00343. [Online]. Available:
[5] K. Bonawitz et al., “Towards federated learning at scale: http://arxiv.org/abs/1809.00343
System design,” 2019, arXiv:1902.01046. [Online]. Available:
[28] G. Zhu, Y. Wang, and K. Huang, “Broadband analog aggregation
http://arxiv.org/abs/1902.01046
for low-latency federated edge learning (extended version),” 2018,
[6] W. Saad, M. Bennis, and M. Chen, “A vision of 6G wireless systems: arXiv:1812.11494. [Online]. Available: http://arxiv.org/abs/1812.11494
Applications, trends, technologies, and open research problems,” IEEE
[29] T. T. Vu, D. T. Ngo, N. H. Tran, H. Q. Ngo, M. N. Dao,
Netw., vol. 34, no. 3, pp. 134–142, May 2020.
and R. H. Middleton, “Cell-free massive MIMO for wireless
[7] J. Park, S. Samarakoon, M. Bennis, and M. Debbah, “Wireless network
federated learning,” 2019, arXiv:1909.12567. [Online]. Available:
intelligence at the edge,” Proc. IEEE, vol. 107, no. 11, pp. 2204–2239,
http://arxiv.org/abs/1909.12567
Nov. 2019.
[8] M. Chen, O. Semiari, W. Saad, X. Liu, and C. Yin, “Federated echo state [30] Y. Sun, S. Zhou, and D. Gündüz, “Energy-aware analog aggregation
learning for minimizing breaks in presence in wireless virtual reality for federated learning with redundant data,” 2019, arXiv:1911.00188.
networks,” IEEE Trans. Wireless Commun., vol. 19, no. 1, pp. 177–191, [Online]. Available: http://arxiv.org/abs/1911.00188
Jan. 2020. [31] K. Yang, T. Jiang, Y. Shi, and Z. Ding, “Federated learning via
[9] H. Tembine, Distributed Strategic Learning for Wireless Engineers. over-the-air computation,” 2018, arXiv:1812.11750. [Online]. Available:
Boca Raton, FL, USA: CRC Press, 2018. http://arxiv.org/abs/1812.11750
[10] S. Samarakoon, M. Bennis, W. Saad, and M. Debbah, “Dis- [32] Q. Zeng, Y. Du, K. K. Leung, and K. Huang, “Energy-efficient radio
tributed federated learning for ultra-reliable low-latency vehicu- resource allocation for federated edge learning,” 2019 arXiv:1907.06040.
lar communications,” 2018, arXiv:1807.08127. [Online]. Available: [Online]. Available: http://arxiv.org/abs/1907.06040
http://arxiv.org/abs/1807.08127 [33] M. Chen, Z. Yang, W. Saad, C. Yin, H. V. Poor, and S. Cui, “A joint
[11] D. Gündüz, P. de Kerret, N. Sidiroupoulos, D. Gesbert, C. Murthy, and learning and communications framework for federated learning over
M. van der Schaar, “Machine learning in the air,” IEEE J. Sel. Areas wireless networks,” IEEE Trans. Wireless Commun., 2020.
Commun., vol. 37, no. 10, pp. 2184–2199, Oct. 2019. [34] N. H. Tran, W. Bao, A. Zomaya, M. N. H. Nguyen, and C. S. Hong,
[12] M. Chen, U. Challita, W. Saad, C. Yin, and M. Debbah, “Artificial “Federated learning over wireless networks: Optimization model design
neural networks-based machine learning for wireless networks: A tuto- and analysis,” in Proc. IEEE Conf. Comput. Commun. (INFOCOM),
rial,” IEEE Commun. Surveys Tuts., vol. 21, no. 4, pp. 3039–3071, Paris, France, Apr. 2019, pp. 1387–1395.
4th Quart., 2019. [35] Y. Mao, J. Zhang, and K. B. Letaief, “Dynamic computation offloading
[13] M. S. H. Abad, E. Ozfatura, D. Gündüz, and O. Ercetin, “Hierarchi- for mobile-edge computing with energy harvesting devices,” IEEE J. Sel.
cal federated learning across heterogeneous cellular networks,” 2019, Areas Commun., vol. 34, no. 12, pp. 3590–3605, Dec. 2016.
arXiv:1909.02362. [Online]. Available: http://arxiv.org/abs/1909.02362 [36] O. Shamir, N. Srebro, and T. Zhang, “Communication-efficient dis-
[14] T. Nishio and R. Yonetani, “Client selection for federated learning with tributed optimization using an approximate newton-type method,”
heterogeneous resources in mobile edge,” in Proc. IEEE Int. Conf. in Proc. Int. Conf. Mach. Learn., Beijing, China, Jun. 2014,
Commun. (ICC), Shanghai, China, May 2019, pp. 1–7. pp. 1000–1008.
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.
YANG et al.: ENERGY EFFICIENT FEDERATED LEARNING OVER WIRELESS COMMUNICATION NETWORKS 1949
[37] Z. Allen-Zhu, “Natasha 2: Faster non-convex optimization than SGD,” Walid Saad (Fellow, IEEE) received the Ph.D.
in Proc. Adv. Neural Inf. Process. Syst., 2018, pp. 2675–2686. degree from the University of Oslo in 2010. He is
[38] W. Dinkelbach, “On nonlinear fractional programming,” Manage. Sci., currently a Professor with the Department of Elec-
vol. 13, no. 7, pp. 492–498, Mar. 1967. trical and Computer Engineering, Virginia Tech,
[39] K. Buza, “Feedback prediction for blogs,” in Proc. Data Anal., where he leads the Network sciEnce, Wireless, and
Mach. Learn. Knowl. Discovery. Cham, Switzerland: Springer, 2014, Security (NEWS) laboratory. His research inter-
pp. 145–152. ests include wireless networks, machine learning,
[40] C. J. Shallue, J. Lee, J. Antognini, J. Sohl-Dickstein, R. Frostig, game theory, security, unmanned aerial vehicles,
and G. E. Dahl, “Measuring the effects of data parallelism on cyber-physical systems, and network science. He is a
neural network training,” 2018, arXiv:1811.03600. [Online]. Available: Fellow of a Distinguished Lecturer of IEEE. He was
http://arxiv.org/abs/1811.03600 also a recipient of the NSF CAREER Award in 2013,
[41] T. Lin, S. U. Stich, K. K. Patel, and M. Jaggi, “Don’t use large mini- the AFOSR Summer Faculty Fellowship in 2014, and the Young Investigator
batches, use local SGD,” 2018, arXiv:1808.07217. [Online]. Available: Award from the Office of Naval Research (ONR) in 2015. He was the
http://arxiv.org/abs/1808.07217 author/coauthor of nine conference best paper awards at WiOpt in 2009,
ICIMP in 2010, IEEE WCNC in 2012, IEEE PIMRC in 2015, IEEE
SmartGridComm in 2015, EuCNC in 2017, IEEE GLOBECOM in 2018, IFIP
NTMS in 2019, and IEEE ICC in 2020. He was a recipient of the 2015 Fred
W. Ellersick Prize from the IEEE Communications Society, of the 2017 IEEE
ComSoc Best Young Professional in Academia Award, of the 2018 IEEE
ComSoc Radio Communications Committee Early Achievement Award, and
of the 2019 IEEE ComSoc Communication Theory Technical Committee.
He was also a coauthor of the 2019 IEEE Communications Society Young
Author Best Paper. From 2015 to 2017, he was named the Stephen O. Lane
Zhaohui Yang (Member, IEEE) received the B.S. Junior Faculty Fellow at Virginia Tech in 2017, and he was named College
degree in information science and engineering from of Engineering Faculty Fellow. He received the Dean’s award for Research
Chien-Shiung Wu Honors College, Southeast Uni- Excellence from Virginia Tech in 2019. He currently serves as an Editor
versity, Nanjing, China, in June 2014, and the Ph.D. for the IEEE T RANSACTIONS ON W IRELESS C OMMUNICATIONS, the IEEE
degree in communication and information system T RANSACTIONS ON M OBILE C OMPUTING, and the IEEE T RANSACTIONS
with the National Mobile Communications Research ON C OGNITIVE C OMMUNICATIONS AND N ETWORKING . He is also the
Laboratory, Southeast University, in May 2018. Editor-at-Large for the IEEE T RANSACTIONS ON C OMMUNICATIONS.
He is currently a Post-Doctoral Research Associate Choong Seon Hong (Senior Member, IEEE)
with the Center for Telecommunications Research, received the B.S. and M.S. degrees in electronic
Department of Informatics, King’s College London, engineering from Kyung Hee University, Seoul,
U.K., where he has been involved in the EPSRC South Korea, in 1983 and 1985, respectively, and the
SENSE Project. He has guest edited a 2020 IEEE Feature topic of IEEE Ph.D. degree from Keio University, Minato, Japan,
Communications Magazine on Communication Technologies for Efficient in 1997. In 1988, he joined Korea Telecom, where
Edge Learning. He was a TPC member of IEEE ICC from 2015 to 2020 and he worked on broadband networks, as a Member of
GLOBECOM from 2017 to 2020. He was an Exemplary Reviewer for Technical Staff. In September 1993, he joined Keio
the IEEE T RANSACTIONS ON C OMMUNICATIONS in 2019. His research University. He worked for the Telecommunications
interests include federated learning, reconfigurable intelligent surface, UAV, Network Laboratory, Korea Telecom, as a Senior
MEC, energy harvesting, and NOMA. He serves as the Co-Chair for the Member of Technical Staff, and the Director of the
IEEE International Conference on Communications (ICC), 1st Workshop on Networking Research Team, till August 1999. Since September 1999, he has
Edge Machine Learning for 5G Mobile Networks and Beyond, June 2020, been a Professor with the Department of Computer Science and Engineering,
IEEE International Symposium on Personal, Indoor and Mobile Radio Com- Kyung Hee University. His research interests include future Internet, ad hoc
munication (PIMRC) Workshop on Rate-Splitting and Robust Interference networks, network management, and network security. He is a member of
Management for Beyond 5G, August 2020. ACM, IEICE, IPSJ, KIISE, KICS, KIPS, and OSIA. He has served as the
General Chair, a TPC Chair/Member, or an Organizing Committee Member of
international conferences, such as NOMS, IM, APNOMS, E2EMON, CCNC,
ADSN, ICPP, DIM, WISA, BcN, TINA, SAINT, and ICOIN. He was an
Associate Editor of the IEEE T RANSACTIONS ON N ETWORK AND S ERVICE
M ANAGEMENT, the International Journal of Network Management, and the
Journal of Communications and Networks; and an Associate Technical Editor
of IEEE Communications Magazine.
Mohammad Shikh-Bahaei (Senior Member, IEEE)
Mingzhe Chen (Member, IEEE) received the Ph.D. received the B.Sc. degree from the University of
degree from the Beijing University of Posts and Tehran, Tehran, Iran, in 1992, the M.Sc. degree
Telecommunications, Beijing, China, in 2019. From from the Sharif University of Technology, Tehran,
2016 to 2019, he was a Visiting Researcher with the in 1994, and the Ph.D. degree from King’s College
Department of Electrical and Computer Engineer- London, U.K., in 2000. He has worked for two
ing, Virginia Tech. He is currently a Post-Doctoral start-up companies and for National Semiconductor
Researcher with the Electrical Engineering Depart- Corporation, Santa Clara, CA, USA (now part of
ment, Princeton University, and at The Chinese Uni- Texas Instruments Incorporated), on the design of
versity of Hong Kong Shenzhen, China. His research third-generation (3G) mobile handsets, for which he
interests include federated learning, reinforcement has been awarded three U.S. patents as an inventor
learning, virtual reality, unmanned aerial vehicles, and a co-inventor. In 2002, he joined King’s College London, as a Lecturer,
and wireless networks. He was a recipient of the IEEE International Confer- where he is currently a Full Professor of Telecommunications with the
ence on Communications (ICC) 2020 Best Paper Award, and an Exemplary Center for Telecommunications Research, Department of Engineering. He has
Reviewer of the IEEE T RANSACTIONS ON W IRELESS C OMMUNICATIONS authored or coauthored numerous journal articles and conference papers.
in 2018 and the IEEE T RANSACTIONS ON C OMMUNICATIONS in 2018 and He has been engaged in research in the area of wireless communications and
2019. He has served as the Co-Chair for workshops on edge learning and signal processing for 25 years both in academic and industrial organizations.
wireless communications in several conferences, including IEEE ICC, IEEE His research interests include the fields of learning-based resource allocation
Global Communications Conference, and IEEE Wireless Communications and for multimedia applications over heterogeneous communication networks, full
Networking Conference. He currently serves as an Associate Editor for Fron- duplex, and UAV communications, and secure communication over wireless
tiers in Communications and Networks, Specialty Section on Data Science for networks. He was the Founder and Chair of the Wireless Advanced (formerly
Communications and a Guest Editor for the IEEE J OURNAL ON S ELECTED SPWC) Annual International Conference from 2003 to 2012. He was a
A REAS IN C OMMUNICATIONS (JSAC) Special Issue on Distributed Learning recipient of the overall King’s College London Excellence in Supervisory
over Wireless Edge Networks. Award in 2014.
Authorized licensed use limited to: Norges Teknisk-Naturvitenskapelige Universitet. Downloaded on April 11,2022 at 11:09:14 UTC from IEEE Xplore. Restrictions apply.