An Energy-Effcient Framework for Data
Aggregation in Wireless Sensor Networks based
on Distributed Source Coding
Tallal Osama EI-Shabrawy
Information Engineering & Technology
German University in Cairo
Cairo, Egypt
tallal.el-shabrawy@guc.edu.eg
Abstract-This paper presents a data aggregation and
forwarding framework in wireless sensor networks (WSNs) that will
help in reducing energy consumption and hence prolong network
lifetime. This approach is based on the fact that WSNs usually
contain a large number of sensor nodes typically with highly
correlated data readings. The proposal is to deploy distributed
source coding (DSC) in compressing data messages to reduce
transmission energy requirements along with avoiding data
redundancy. Specifcally, a DSC construction is proposed to
determine the number of bits needed to encode a data message by a
sensor node relative to its correlated neighbors without exchanging
excessive communication messages among them.
Keword-Ener Consumton; Data aggregation and
forwardng; Distributed Source Coding; Wreless Sensor Network
I. INTRODUCTION
The importance of wireless sensor networks (WSNs) over the
past few years have increased with the continuous development
of sensors technology. While sensors might seem tiny and cheap,
they are powerfl devices with computing and communication
facilities that are used in many applications such as militarily,
health, industry and surveillance. Nonetheless, sensors are
energy constrained, which makes energy consumption a critical
matter for WSNs. Experiencing excessive large communication
messages will eventually drain network energy. As a result, this
might shorten the lifetime of WSNs. Hence, these concers
might act as obstacles to the expected huge growth of WSN in
the near fture.
Thus, many protocols have been proposed lately to ofer a
reliable practice for maximizing network lifetime via reducing
energy consumption. Reducing energy consumption can be done
through diferent approaches. One approach is through exploiting
the network topology as proposed in [3, 7], or through routing as
done by [4 - 5], or by reducing data size and number of
transmissions as in [8, 9], respectively, or fnally by data
aggregation as explained in [6, 10].
The proposed famework, on the other hand, proposes a data
aggregation protocol based on encoding correlated data at
sensors before forwarding them to the sink. This approach aims
to reduce energy consumption and hence maximize network
lifetime. The establishment of this approach is based on
deploying distributed source coding (DSC) [1, 2] for encoding
aggregated data at sensor node without the need for accessing
neighbor's correlated data. To determine the number of encoding
bits, a simple correlation tracking technique is employed.
The famework was performed and tested in a clustering-based
environment that exploits a multiple-hop routing for aggregating
978-1-61284-185-4/11/$26.00 ©2011 IEEE 56
Nora Mohamed Mounir
Media Engineering & Technology
German University in Cairo
Cairo, Egypt
nora.mohamed@guc.edu.eg
data along the path between two end nodes. This helped in
reducing number of data transmissions and enhancing
redundancy avoidance, hence reducing energy consumption
along with enhancing overall system effciency. The paper is
organized as follows. Section II provides technical background
about DSC. The famework is presented in section III. Section IV
presents simulation results and Section V concludes the work.
II. BACKGROUND
Distributed source coding as lossless data compression was
frst introduced by Slepian and Wolf in 1973 [1]. Slepian and
Wolf theorem states that if a pair of correlated discrete random
information sources is uniformly distributed, then they can be
encoded separately as efcient as the case where the two sources
being encoded together.
To explain Slepian-Wolf theorem, consider X and Y to be a
pair of correlated discrete random information sources. From
Shannon's source coding theory [13], the two sources can be
compressed sufciently if they are encoded together by a rate
given by their joint entropy H( 1 of X and Y. In such case, Y
is compressed frst into H( bits/sample, and X is then
compressed into H ( 1 bits/sample where both the encoder and
the decoder have a complete knowledge of Y. According to
Slepian-Wolf, X and Y can be separately compressed providing
the same compression effciency as if they were compressed
together. This is achievable through compressing X and Y at rates
Rx and Ry respectively such that [1]
Rx � H(X I Y),
Ry � H(Y I X),and
Rx + R
y
� H(X,Y)
(1)
While proving their theorem, Slepian-Wolf used the random
binning concept. Binning refers to dividing all possible outcomes
for decoding X into sets of pairs. Then, the transmitted bits fom
X present the index of the set which X belongs to.
To see how this works, consider the following example, which
was demonstrated in [12], where X and Y are two discrete
random sets of 3-bits data whose correlation can be described as
followS:d
H(
X,Y)�I, where d
H(
X
,Y)
denotes the Hamming
distance between the two sets. Then their entropies H(X and
H( equal to 3-bits. For encoding X without accessing Y by the
encoder but can be accessed by the decoder, all possible
outcomes for decoding X are divided into four sets of pairs
(bins). These four bins are Zoo ={OOO) 11 }, Z
O
I
={OO 1) 1O } .
Z
I
O
={OIO)OI}, Zll ={OIl)OO). Providing that the Hamming
distance between those pairs is the maximum (which equal to 3