Apnoms 2006
Apnoms 2006
Editorial Board
David Hutchison
Lancaster University, UK
Takeo Kanade
Carnegie Mellon University, Pittsburgh, PA, USA
Josef Kittler
University of Surrey, Guildford, UK
Jon M. Kleinberg
Cornell University, Ithaca, NY, USA
Friedemann Mattern
ETH Zurich, Switzerland
John C. Mitchell
Stanford University, CA, USA
Moni Naor
Weizmann Institute of Science, Rehovot, Israel
Oscar Nierstrasz
University of Bern, Switzerland
C. Pandu Rangan
Indian Institute of Technology, Madras, India
Bernhard Steffen
University of Dortmund, Germany
Madhu Sudan
Massachusetts Institute of Technology, MA, USA
Demetri Terzopoulos
University of California, Los Angeles, CA, USA
Doug Tygar
University of California, Berkeley, CA, USA
Moshe Y. Vardi
Rice University, Houston, TX, USA
Gerhard Weikum
Max-Planck Institute of Computer Science, Saarbruecken, Germany
Young-Tak Kim Makoto Takano (Eds.)
Management of
Convergence Networks
and Services
13
Volume Editors
Young-Tak Kim
Yeungnam University
School of Electronics Engineering and Computer Science
214-1, Dae-Dong, Kyungsan-Si, Kyungbook, 721-749, Korea
E-mail: [email protected]
Makoto Takano
NTT West R&D Center
IT Governance Promoting Group
6-2-82, Shimaya, Konohana-ku, Osaka, 554-0024, Japan
E-mail: [email protected]
ISSN 0302-9743
ISBN-10 3-540-45776-3 Springer Berlin Heidelberg New York
ISBN-13 978-3-540-45776-3 Springer Berlin Heidelberg New York
This work is subject to copyright. All rights are reserved, whether the whole or part of the material is
concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting,
reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication
or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965,
in its current version, and permission for use must always be obtained from Springer. Violations are liable
to prosecution under the German Copyright Law.
Springer is a part of Springer Science+Business Media
springer.com
© Springer-Verlag Berlin Heidelberg 2006
Printed in Germany
Typesetting: Camera-ready by author, data conversion by Scientific Publishing Services, Chennai, India
Printed on acid-free paper SPIN: 11876601 06/3142 543210
Preface
General Chair
James Hong, POSTECH, Korea
Vice Chair
TPC Co-chairs
Tutorial Co-chairs
Keynotes Chair
DEP Chair
Exhibition Co-chairs
Poster Co-chairs
Publicity Co-chairs
Gilhaeng Lee, ETRI, Korea
Choong Seon Hong, KHU, Korea
Qinzheng Kong, HP APJ, Australia
Kiminori Sugauchi, Hitachi, Japan
Finance Co-chairs
Hong-Taek Ju, Keimyung University, Korea
Kohei Iseda, Fujitsu Labs., Japan
Publication Chair
YoungJoon Lee, KNUE, Korea
Secretaries
Jae-Oh Lee, KUT, Korea
Hideo Imanaka, NTT, Japan
International Liaisons
Takeo Hamada, Fujitsu Labs of America, USA
Raouf Boutaba, University of Waterloo, Canada
Carlos Westphall, Santa Catalina Federal University, Brazil
Hiroyuki Okazaki, NEC Europe Ltd., Germany
Rajan Shankaran, Macquarie University, Australia
Alpna J. Doshi, Satyam Computer Services, India
Teerapat Sanguankotchakorn, AIT, Thailand
Ryoichi Komiya, Multimedia University, Malaysia
Victor WJ Chiu, Chunghwa Telecom, Taiwan
Yan Ma, Beijing University of Posts and Telecommunications, China
Advisory Board
Young-Hyun Cho, KTH, Korea
Young-Myoung Kim, KT, Korea
Graham Chen, EPAC Tech, Australia
Makoto Yoshida, University of Tokyo, Japan
Masayoshi Ejiri, Fujitsu, Japan
Doug Zuckerman, Telcordia, USA
Organization IX
Standing Committee
Nobuo Fuji, NTT, Japan
Hiroshi Kuriyama, NEC, Japan
James W. Hong, POSTECH, Korea
Kyung-Hyu Lee, ETRI, Korea
Seong-Beom Kim, KT, Korea
Yoshiaki Tanaka, Waseda University, Japan
Additional Reviewers
1 Introduction
A wireless ad hoc network consists of a group of mobile nodes without the support of
any infrastructure. Such a network is expected to support advanced applications such
as communications in emergency disaster management, video conferencing in a
workshop or seminar, communications in a battlefield. This class of mission-critical
applications demands a certain level of quality of services (QoS) for proper opera-
tions. Also due to the distributed nature of these networks providing a fair access of
resource to multiple contending nodes is an important design issue.
Fairness is an important criterion of resource sharing in the best effort Internet, es-
pecially when there is a competition for the resource among the nodes due to unsatis-
fied demands. In fair scheduling each flow f is allowed to share a certain percentage
of link capacity based on its flow weight indicated as wf. Let Wf (t1, t2) and Wg (t1, t2)
denote the aggregate resources received by flows f and g respectively in time interval
[t1, t2] and wf and wg are the flow weights of the flows f and g respectively. The
allocation is ideally fair if it satisfies (1)
*
This work was supported by MIC and ITRC Project.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 1 – 10, 2006.
© Springer-Verlag Berlin Heidelberg 2006
2 M.M. Alam, Md. Mamun-or-Rashid, and C.S. Hong
W f ( t1 , t 2 ) W g ( t1, t 2 )
− = 0 (1)
w f wg
However, most of the research works assumed error free wireless link which is not
realistic. In the wireless environment, a packet flow may experience channel error and
the transmission may not be successful. Thus the bursty and location dependent error
in wireless link may make the existing fair scheduling algorithm inapplicable. There-
fore, the goal of ad hoc network fair scheduling is to make short burst of location-
dependent channel error transparent to users by a dynamic reassignment of channel
allocation over small timescales [14]. Specifically, a backlogged flow f that perceives
a channel error during a time window [t1, t2] is compensated over a later time window
[t1', t2'] when f perceives a clean channel. Compensation for f involves granting addi-
tional channel access to f during [t1', t2'] in order to make up lost channel access dur-
ing [t1, t2], and this additional channel access is granted to f at the expense of flows
that were granted additional channel access during [t1, t2].
Providing QoS in Wireless ad hoc networks is a new area of research. Existing
works focuse mainly on QoS routing which finds a path to meet the desired service
requirements of a flow. In this paper we consider a mix of guaranteed and best effort
flows and investigate fair queueing with QoS support for the network. The goal is to
guarantee the minimum bandwidth requirements of guaranteed flows and to ensure a
fair share of residual bandwidth to all flows.
In this paper we focus on the fair scheduling issues in a hoc network in the pres-
ence of channel errors. We develop a fairness model for wireless ad hoc fair schedul-
ing to deal with channel error. We also implement the model in a distributed manner
by localizing the global information required by the nodes.
The rest of the paper is organized as follows. Section 2 describes related works. In
Section 3 we explain the network model and problem specifications. Section 4 de-
scribes the proposed mechanism and is followed by the details of the implementation
of the proposed mechanism in section 5. Section 6 presents the simulation and results.
We conclude in section 7 by conclusion and future works.
2 Related Works
Fair queueing has been a popular paradigm for providing fairness, minimum through-
put assurance and guaranteed delay in wired network [1], and in packet cellular net-
works [2] – [4]. Recently some techniques have been proposed to incorporated fair
queueing in shared channel, multihop wireless networks [5] – [7]. Also, providing
QoS in wireless ad hoc networks is a new area of research. Some of the research
works also incorporated both QoS and fair queueing in ad hoc networks. Both QoS
guarantee and fair queueing in ad hoc networks have been proposed in [8] and [9].
Also some of the works provide fairness with error compensation, e.g., [13], [14],
[15] and most of these are proposed based on the support of base stations.
In [13], channel-condition independent packet fair queueing (CIF-Q) is proposed.
Each session is associated with a parameter called lag to indicate whether the session
should be compensated. If a session is not leading and the channel is error free at
its scheduled time, its head-of-line packet is transmitted; otherwise, the time slot is
QoS-Aware Fair Scheduling in Wireless Ad Hoc Networks with Link Errors 3
released to other sessions. The problem with CIF-Q is that the leading sessions are not
allowed to terminate unless all their leads have been paid back, regardless of whether
such terminations are caused by broken routes. This property makes CIF-Q an inade-
quate solution for ad hoc networks, because a connection may be broken.
The Idealized Wireless Fair-Queueing (IWFQ), and the Wireless Packet Schedul-
ing protocol (WPS) are proposed in [14]. In this paper the base station calculates the
number of slots per frame, each flow can use, based on the weight of the flows. If a
flow experiences channel error at its allocated time, the base station tries to find a
flow that can exchange its slot with the flow within the frame; and the error session of
the flow is compensated at a later frame. The mechanism is not suitable for ad hoc
networks, as it requires the presence of a base station.
In [15], QoS-aware fair scheduling is proposed for mobile ad hoc networks in the
presence of error; and based on the channel condition estimation, a flow may either
transmit or give up its allocation to others. When this flow perceives an error free
channel it will have packets with smaller service tag than that of its neighbors’ and its
packets will be transmitted first. The problem with this protocol is that after recovery
from error a flow exclusively accesses the channel until its virtual time catches up
with other flows. Also the service release of a leading flow is not graceful.
iv) The extra service releases by a leading flow should be graceful to guarantee
short-term fairness.
We consider a multi-hop flow consists of number of single hop flows and the flow
weight is not fixed for the path; every forwarding node assigns a different weight.
Over time, based on the current flows within two-hop neighbors, the weight may
change, otherwise either the network utilization or the fairness will be poor.
3) Tagging Operations: For each flow f we use the SFQ [10] algorithm to assign
tags for the arriving packets: a start tag and a finish tag.
4) Path Registration: To provide guaranteed service a routing protocol should find
a feasible path. AODV [11] is one of the most widely used table-based and reactive
routing protocols. But AODV is designed to find a feasible route only. Therefore, to
support QoS we need to modify AODV. To ensure QoS, AODV is modified to sup-
port two types of schemes: admission scheme and adaptive scheme. In admission
scheme a feasible path should provide the required minimum bandwidth, while in
adaptive feedback scheme the source is informed about the minimum available band-
width so that the source can adapt its transmission speed.
QoS-Aware Fair Scheduling in Wireless Ad Hoc Networks with Link Errors 5
To initiate QoS-aware routing discovery, the source host sends a RREQ packet
whose header is changed to <model-flag, required bandwidth, min-bandwidth, AODV
RREQ header>. The model-flag indicates whether the source is using the admission
scheme or the adaptive feedback scheme. When an intermediate host receives the
RREQ packet, it first calculates its residual bandwidth. If the model-flag is set to
admission scheme, the host compares its residual bandwidth with the minimum re-
quested bandwidth. If its residual bandwidth is greater than the minimum bandwidth,
it forwards this RREQ. Otherwise, it discards this RREQ. If the model-flag is adap-
tive, the host compares its residual bandwidth with the min-bandwidth field in the
RREQ. If its residual bandwidth is greater than the min-bandwidth, it forwards the
RREQ. Otherwise, it updates the min-bandwidth value using its residual bandwidth.
Finally the forwarding node temporarily stores the flow. When a node forwards a
RREP message it assigns a flow-weight to the flow and stores the flow information.
5) Channel Error Prediction: Perfect channel-dependent scheduling is only possi-
ble if a node has accurate information about the channel state. The location-dependent
nature of channel error requires each node to monitor its channel state continuously,
based on which the node may predict its future channel state. To do this, each node
periodically measures the signal-to-noise ratio (SNR) of the channel. If the SNR value
falls below a predefined threshold, the channel is considered as error-prone.
6) Lead and Lag Model: The purpose of the lead and lag model is to determine
how much additional service a lagging flow is entitled to receive in the future in order
to compensate service lost in the past and how much service a leading flow should
relinquish in the future in order to give up additional services received in the past.
Also, the lag of a lagging flow is incremented on a lost time slot only if another flow
received its service and is ready to release this service.
To represent the amount of lead and lag service of a flow we use a parameter called
lag. The value of lag of a lagging flow is positive, it is negative for a leading flow and
zero otherwise.
When a node fi perceives an erroneous wireless link and it is the turn for one of its
flow to transmit, the node does not transmit. Instead one of its neighbors, fj, which
perceives a good channel and has the immediate higher service tag (lagging flows get
higher preference than leading flows) will transmit. Both of the flows will either ini-
tialize (if they just become lagging and leading) or update their lag value.
7) Compensation Model: The compensation model is the key component of ad hoc
network fair scheduling algorithm in the presence of errors. It determines how lagging
flows receive extra service to make up their lag and leading flows give up their lead to
relinquish extra service. The compensation model should take into account the fol-
lowing two conditions to ensure short-term fairness:
i) The service releases by a leading flow should not be more than a certain fraction
of its received service.
ii) The lagging flows should receive the extra service in proportion to their flow
weights. That is, flow with largest lag should not receive more service irrespective to
its guarantee rate.
To achieve graceful degradation of service of a leading flow, it relinquishes a frac-
tion of services allocated to the flow. We define a system parameter Į (0 Į 1) to
control the minimal fraction of service retained by a leading session. That is, a leading
flow can give up at most (1 – Į) amount of its service to compensate for lagging
6 M.M. Alam, Md. Mamun-or-Rashid, and C.S. Hong
flows. Each leading flow, fi, is associated with another parameter, called normalized
service received by the leading flow, si. When a flow becomes leading, si is initialized
to Įvi (vi is the virtual time of fi) and si is updated whenever fi is served.
On the other hand, lagging flows should have higher priority to receive extra ser-
vice to compensate their loss. And such compensation is possible because leading
flows give up their leads. To provide short-term fairness we distribute these additional
services among lagging flows in proportion to the lagging flows weights. To accom-
plish this we use the compensation virtual time, ci, which keeps the track of normal-
ized amount of compensation services received by a flow while it is lagging. When
flow fi becomes lagging ci is initialized according to (2)
ª º
c i = « max( c , min { f | lag > 0} » (2)
i k k
¬« k∈A »¼
8) Scheduling Mechanism: For scheduling the flows we use a table driven, back-
off based approach which uses local information and local computation only. With the
tagging operation and a method of exchanging tags, each node has the knowledge of
its local neighborhood. These tags are stored in tables and are ordered so that each
node can learn whether that node itself has the minimum tag and therefore has to
transmit next. The basic scheduling mechanism is as follows:
i) When a node finds one of its flows fi has the minimum service tag and the flow is
not leading or leading but did not receive a minimum fraction of service, then the
packet at its queue is transmitted. However, if fi is leading and already receives mini-
mum service, it does not transmit its packet to give up its lead to a lagging flow. If
there is any lagging flow within two-hop neighbors of the flow, then the flow fj, which
has the minimum compensation tag among the lagging flows can transmit its packet,
otherwise fi transmits its packet.
ii) After the packet to transmit has been determined, the virtual time of fi is up-
dated. If fi is leading but receives services due to graceful degradation, si is updated to
si + Lp. If fj is served and the overhead is charged to fi where (i j), then the lag val-
ues are update by lagj = lagj – Lp and lagi = lagi + Lp.
iii) When a flow just becomes lagging its compensation_tag, ci is initialized by (2)
and updated every time a lagging flow gets compensation service by ci =ci + Lp / wi.
When flow fi becomes leading its si is initialized by si = Įvi and updated every time
it gets service due to graceful degradation by si = si + Lp / wi.
9) Table Update: Whenever a node hears a new service tag for any flow on its ta-
ble or a new flow, it updates the table entry for that flow or adds this flow information
on its table. Whenever any node transmits a head-of-line packet for a flow, it updates
that flow’s service tag and compensation tag in the table entry.
neighbors have the information regarding flow fi, it is known to all that fi should relin-
quish its extra service in this slot. Therefore, the next flow will be selected from one
of the lagging flows (if there is any) and determined based on the compensation_tag
instead of service_tag. Therefore, once a leading flow has the minimum service tag
which has to release its extra services, backoff values are assigned only to lagging
flows and as a result one of the lagging flows is scheduled to transmit its packet.
5) Propagation of Updated Service Tag: In order to propagate a flow’s service tag
to all its one-hop neighbors in the node graph and reduce the chance of information
loss due to collisions during the propagation, we attach the service_tag, compensa-
tion_tag and si for flow fi in all four packets RTS, CTS, DS and ACK. However, we
do not use the updated tags for flow fi in RTS and CTS packets, since RTS and CTS
do not ensure a successful transmission. When the handshake of RTS and CTS is
completed, we attach the updated flow tag in DS and ACK, to inform neighboring
nodes about the new updated information of the current transmitting flow fi.
f0
f1 f2 f3
Figure 2 and Figure 3 show the simulation result of our proposed mechanism. In
Figure 2, we consider all the flows are experiencing error-free wireless channel.
In Figure 2(a), all the flows are best-effort flows and have the same weight. So, all
flows have got the equal share of the bandwidth. In Figure 2(b), flow f2 is guaranteed
flow and all other flows are best-effort flows. Flow f2 has the minimum bandwidth
requirement of 0.2C, where C is the link bandwidth. So the flow weights are w2 = 2
and w0 = w1 = w3 = 1. Accordingly, flow f2 got double share than the other flows.
In Figure 3, flow f0 is experiencing an erroneous channel during time interval 4 to 9
seconds and so in this period flow f0 does not receive any service. Figure 3(a) shows
that, at that time f1 receives the service (lost service of f0) and as f1 and f3 can transmit
QoS-Aware Fair Scheduling in Wireless Ad Hoc Networks with Link Errors 9
(a) (b)
Fig. 2. Simulation results with error-free channel (a) All flows are best-effort (b) Flow f2 is
guaranteed, all others are best-effort
(a) (b)
Fig. 3. Simulation results with error in wireless channel (a) all flows are best-effort and (b) flow
f2 is guaranteed flow and all the other flows are best-effort
simultaneously, so f3 also receives extra service. As the graph shows the lost service
of f0 is compensated at time 11 and 13 seconds, which justify our proposed algorithm.
In Figure 3(b), f2 is guaranteed flow and all other flows are best-effort flows. And
flow f0 experiences erroneous channel during time interval 4 to 9 seconds and later its
service is paid back by flow f1, f2 and f3 those who received the extra service earlier.
7 Conclusions
In this paper, we proposed a distributed fair scheduling algorithm for providing
scheduling service in wireless ad hoc networks with errors in wireless link. Our pro-
posed mechanism provides a compensation model for flows that experience an erro-
neous channel, and allows the flows to get their lost service at a later time. It also
ensures that a leading flow can give up its extra service in a graceful way. Also we
considered the presence of both QoS and best-effort flows simultaneously in the net-
works. Our proposed mechanism satisfies the minimum bandwidth requirement of
guaranteed flows and provides a fair share of the residual bandwidth to all flows. As a
future work, we like to apply our proposed algorithm to mobile ad hoc networks.
10 M.M. Alam, Md. Mamun-or-Rashid, and C.S. Hong
References
1. A. Demers, S. Keshav, and S. Shenker: Analysis and simulation of a fair queueing algo-
rithm. ACM SIGCOMM, (1989) 1–12
2. S. Lu, V. Bharghavan, and R. Srikant: Fair scheduling in wireless packet networks.
IEEE/ACM Transaction On Networking Vol. 7 (1999) 473–489
3. T. S. Ng, I. Stoica, and H. Zhang: Packet fair queueing algorithms for wireless networks
with location-dependent errors. IEEE INFOCOM (1998) 1103–1111
4. P. Ramanathan and P. Agrawal: Adapting packet fair queueing algorithms to wireless net-
works. ACM MOBICOM (1998) 1–9
5. H. Luo and S. Lu: A self-coordinating approach to distributed fair queueing in ad hoc
wireless networks. IEEE INFOCOM (2001) 1370–1379
6. H. Luo and S. Lu: A topology-independent fair queueing model in ad hoc wireless net-
works. IEEE Int. Conf. Network Protocols (2000) 325–335
7. H. L. Chao and W. Liao: Credit-based fair scheduling in wireless ad hoc networks. IEEE
Vehicular Technology Conf. (2002)
8. Jerry Cheng and Songwu Lu: Achieving Delay and Throughput Decoupling in Distributed
Fair Queueing Over Ad Hoc Networks. IEEE ICCCN (2003)
9. H. L. Chao and W. Liao: Fair Scheduling With QoS Support in Wireless Ad Hoc Net-
works. IEEE Trans. on Wireless Comm., vol. 3 (2004)
10. P. Goyal, H.M. Vin and H. Chen: Start-time fair queueing: A scheduling algorithm for in-
tegrated service access. ACM SIGCOMM (1996)
11. C. Perkins, E. Belding-Royer and S. Das: Ad hoc On-Demand Distance Vector (AODV)
Routing. RFC 3561 (2003)
12. V. Bharghavan, A. Demers, S. Shenker, and L. Zhang: MACAW: A Medium Access Pro-
tocol for Wireless LANs. ACM Ann. Conf. Special Interest Group on Data Comm.
(SIGCOMM) (1994)
13. T.S.E. Ng, I. Stoica, H. Zhang: Packet Fair Queueing Algorithms for Wireless Networks
with location-Dependent Errors. IEEE INFOCOM (19998)
14. S. Lu, V. Bharghavan and R Srikant: Fair Scheduling in Wireless Packet Networks.
IEEE/ACM Transaction. On Networking vol. 7 (1999)
15. H. L. Chao and W. Liao: Fair Scheduling in Mobile Ad Hoc Networks with Channel Er-
rors. IEEE Transaction of Wireless Communications, Vol 4 (2005)
16. M. M. Alam, M. M. Rashid and C. S. Hong: Distributed Coordination and Fair Queueing
in Wireless Ad Hoc Networks. Lecture Notes in Computer Science, Vol. 3981 (2006)
Performance Analysis of Service Differentiation for IEEE
802.15.4 Slotted CSMA/CA∗
Meejoung Kim
1 Introduction
The success of wireless sensor networks as a technology depends on the success of
the standardization efforts to unify the market and avoiding the proliferation of
proprietary and incompatible protocols that will limit the size of overall wireless
sensor market. Since IEEE 802.15.4 standard for low data rate wireless personal area
networks (LR-WPAN) supports small, inexpensive, energy efficient devices operating
on battery power that require no infrastructure to operate, it is considered as one of
the technology candidates for wireless sensor networks [1]-[2].
The standard supports two network topologies, the peer-to-peer topology in which
devices can communicate with one another directly as long as they are within the
physical range and the star topology in which devices must communicate through a
central controller device commonly referred as PAN coordinator. It also defines two
channel access mechanisms depending on whether a beacon frame is used to
synchronize communications or not. In the beacon enabled networks, slotted carrier
sense multiple access mechanism with collision avoidance (CSMA/CA) is used and
the slots are aligned with the beacon frame which is sent periodically by the PAN
coordinator. On the other hand, in the non-beacon enabled networks, unslotted
∗
This research was supported by the MIC (Ministry of Information and Communication),
Korea, under the ITRC (Information Technology Research Center) support program super-
vised by the IITA (Institute of Information Technology Assessment).
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 11 – 22, 2006.
© Springer-Verlag Berlin Heidelberg 2006
12 M. Kim
CSMA/CA is used and no beacon frame is used. Since the central device manages the
networks in general and the PAN coordinator can act as the network controller and
the sink to collect data from sensor nodes, for sensor network implementation, the star
topology operating in beacon enabled mode appears to be better suited than the peer-
to-peer topology operating in non-beacon enabled mode.
The related researches deal with the operation of PAN operating under 802.15.4 in
the beacon enabled mode under saturation condition [2-6] and non-saturation [7].
In sensor network, multiple queues cannot be used and only packets with small size
are transmitted dues to the characteristics of the devices. Therefore, the priorities have
to be considered in a slightly different way compare to those using multiple queues in
one node. In this paper, we propose two mechanisms for modified 802.15.4 sensor
networks which provide multi-level differentiated services for each and every device.
To investigate the performance of the proposed mechanisms, mathematical model
which is based on discrete-time Markov chain is presented. By analyzing the Markov
chain, we obtain the probabilities that the medium is idle when the device is sensing,
the probabilities of attempting transmission, throughputs, delays, and drop
probabilities for different priority classes.
The rest of the paper is organized as follows. In Section 2, we propose the
operating mechanisms, and the mathematical model for the proposed mechanisms is
presented in Section 3. Even though several performance measures such as delay and
packet drop probability can be considered, we present the throughput performance
only in Section 4. Section 5 provides the numerical results.
guaranteed
time slot GTS inactive
(GTS)
Q
in each priority q class, in other words n = ¦ nq , where q denotes the priority taking
q =0
integer values in [0,Q]. Throughout the paper, we assume that each and every device
has its own priority which does not change during a superframe and the packet
transmission is completed during the superframe. We assume that packets are
Performance Analysis of Service Differentiation for IEEE 802.15.4 Slotted CSMA/CA 13
generated by Poisson process with arrival rate λq for each device in priority q class
and each device has a finite buffer so that the newly generated packets are dropped
when the buffer is fully occupied. Then the probability that the device does not have
an empty queue after transmit a packet is given by min(1, λq E ( Dq )) , where E ( Dq ) is
the average service time of priority q class. Denote it by ρ q and note that ρ q = 1
implies the saturation mode. Furthermore, we assume that if the transmitted packet
collides, the packet will be dropped and the device will try to transmit a new packet in
the head of the queue if there is a packet waiting for transmission.
Since we conjectured that the parameters, the size of contention window and
backoff exponent, will give different effects on IEEE 802.15.4 wireless sensor
network performances, we consider two service differentiation mechanisms by
varying values of these two parameters. In both differentiation mechanisms, we apply
the scheme which chooses the random backoff counter differentially in each stage. In
other words, during the ith backooff stage, the backoff counter is randomly chosen in
[Wi-1, Wi-1] rather than [0, Wi-1], Wi=2i W0.. This scheme may reduce the collisions by
waiting more time before sensing and gives better chance to access the channel to a
device with early backoff stage. Even though the combined mechanism of the
proposed two mechanisms can be considered, we consider the mechanisms separately
in this paper. In the following subsection, we describe the mechanisms in detail.
Fig. 2 describes the operation of the proposed scheme. Due to the space limitation, we
do not explain the operation step by step.
3 Analytical Model
To analyze the proposed mechanisms, we introduce the following two stochastic
processes for a given device in the priority q class. Let n(q, t) and b(q, t) be the
stochastic processes representing the values of NB and the value of the backoff
counter and/or contention window, respectively, at time t and denote e for empty
buffer. Note that NB represents the backoff stage within the range of [0, m+1], m =
macMaxCSMABackoffs. Then {(n(q, t), b(q, t),e), (n(q, t), b(q, t))} forms a multi-
dimensional Markov process defining the state of the head packet in the queue at the
backoff unit boundaries. Since we are assuming that every and each device has its
own priority which does not change in a superframe and the packet transmission is
completed in the superframe, without loss of generality, each of the processes n(q,
t),b(q, t) can be written simply as n(t), b(t) during the superframe. Then the
corresponding state space is denoted as follows:
Ω = {(0, b0 (t ), e), (n(t ), b(t )), (m + 1, 0) 0 ≤ n(t ) ≤ m, 0 ≤ b0 (t ) ≤ W0 − 1,
−CW [q] ≤ b(t ) ≤ Wi − 1, i = 0," , m, n(t ), b0 (t ), and b(t ) are integers},
Performance Analysis of Service Differentiation for IEEE 802.15.4 Slotted CSMA/CA 15
where W0 = 2BE[q] , Wi = 2iW0, and (0, b0 (t ), e) is the state of the device with empty
buffer. In the proposed model, when the queue is empty the backoff counter is chosen
randomly in [0, W0-1]. If a device is in the state (0,i,e) and the new packet arrives, the
state is changed to (0,i). If the new packet does not generated until the state becomes
(0,0,e), the backoff counter is choose again randomly in [0, W0-1]. In addition, the
state (i, j ), i ∈ [0, m], j ∈ [0,Wi − 1] , and (i, − j ), i ∈ [0, m], j ∈ [1, CW [q]] , denote the
device in the mode of decreasing backoff counter and sensing the channel,
respectively.
Let δ and N (δ ) be the unit of backoff period and the number of packets arrived
during the backoff period, respectively. Then the probability that at least one packet
arrives in δ time interval is given by
− λqδ
Pq ≡ 1 − e . (4)
The state transition diagram of these states is illustrated in Fig 3. For the simplicity
of the notations, we use the transition probabilities P(i1 , j1 | i0 , j0 ) instead of
1 − Pq 1 W0 1 W0 1 W0 1 W0
1 − Pq 1 − Pq 1 − Pq 1 − Pq
0, 0, e 0 , 1, e 0,W 0 − 2, e 0 , W 0 − 1, e
1 W0 Pq Pq Pq
Pq 1 W0 1 W0 1 W0
1 1 1 1 1
0 ,1 0, 2 0 ,W 0 − 2 0,W 0 − 1
0, 0
1− α1 1−α 1−α q (1 − α q +1 )ρq
0, −1 2
0, − 2 0 , − ( q + 1) 0, − q
(1 − α q +1 ) (1 − ρ q )
α1 α 2 α3 αq α q +1
1 (W 1 − W 0 ) 1 (W 1 − W 0 )
1 1 1 1 1
1, 1 1, W 0 − 1 1, W 0 1, W 1 − 1
1, 0
1− α1 1−α 2 1−α q (1 − α q +1 )ρq
1, − 1 1, − 2 1, − ( q + 1) 1, − q
(1 − α q +1 ) (1 − ρ q )
α1 α2 α3 αq α q +1
1 (W 2 − W 1 ) 1 (W 2 − W 1 )
α1 α 2 α3 αq α q +1
1 (W m − W m − 1 ) 1 (W m − W m − 1 )
1 1 1 1 1 1
m ,1 m ,W m −1 −1 m , W m −1 m ,W m − 1
m,0
1− α1 1−α 1−α q (1 − α q +1 )ρq
m ,−1 2
m, −2 m , − ( q − 1) m,−q
α1 α2 α3 αq α q +1 (1 − α q +1 ) (1 − ρ q )
ρq 1− ρq
m + 1, 0
∏ ∏ (1 − τ )
nq −1 1
PI ,n −1 = (1 − τ i ) ni (1 − τ q ) = i
ni
, (14)
i = 0,i ≠ q
1 −τ q i =0
Q Q Q
∏ ∏ (1 − τ )
nq − 2
niτ i (1 − τ i ) ni −1 (1 − τ j ) j (1 − τ q )−1 − τ q (1 − τ q )
n
PS ,n −1 = ¦
i =0 j = 0, j ≠ i i = 0,i ≠ q
i
ni
(15)
and PC ,n −1 = 1 − PI ,n −1 − PS ,n −1 , (16)
where τ q is the probability that a device in the priority q class transmits during a
backoff period and ρ q is the probability of the queue is not empty which is given in
Section 2.
Let TS and TC be the average durations for successful transmission and collision
after the final CCA procedure, respectively. Then TS and TC are given by
TS = TH + Tpacket + twack
s
+ tack and TC = TH + Tpacket + twack
f
, (17)
s f
where TH , Tpacket , twack , tack , and twack are the average durations of transmitting the
header (including MAC header and PHY header), packet transmission, waiting time
for ACK for successful transmission, time for receiving ACK, and waiting time for
ACK for unsuccessful transmission. Then the probability of the channel busy in the
1st CCA procedure which is determined by either successful transmission of one
device or collision out of n-1 devices is given by
PS , n −1TS + PC , n −1TC
α1 = (18)
PI , n −1δ + PS , n −1TS + PC , n −1TC
and subsequently we obtain the following probabilities of channel idle during the jth
CCA procedure:
1 − α j = PI ,n −1 j −1 (1 − α1 ) , j = 1," , CW [q ] + 1 (19)
Performance Analysis of Service Differentiation for IEEE 802.15.4 Slotted CSMA/CA 17
Since all of the states in Ω is positive recurrence and the system is stable,
there exist the stationary probabilities {b0, j0 ,e , bi , j , bm +1,0 : i ∈ [0, m], j0 ∈ [0,W0 − 1],
j ∈ [−CW [q],Wi − 1]} of the discrete-time Markov chain which is defined by
b0, j ,e = lim P (0, b(t ) = j , e) , bi , j = lim P(n(t ) = i, b(t ) = j ) ,
t →∞ t →∞
Wi − j i
bi , j = ACW [ q ]+1b0,0 , i = 1," , m, j = Wi −1 ," ,Wi − 1 , (23)
Wi − Wi −1
j
bi , − j = ∏ (1 − α r ) ACW
i
[ q ] +1b0,0 , i = 1," , m, j = 1," , CW [ q ] , (24)
r =1
1 − (1 − Pq )W0 − j 1
b0, j ,e = ( Bb0,0 + (1 − Pq )b0,0,e ), j = 0," ,W0 − 1 , (25)
Pq W0
CW [ q ] +1 m
1 1
where B is given by B = (1 − ρ q ) ∏ (1 − α r )¦ ACWi
[ q ]+1 + (1 − ρ q ) ACW
m
[ q ] +1 .
W0 r =1 i =0 W0
Furthermore, b0,0 is calculated as
−1
§ 1 − (1 − Pq )W0 · § 1 − (1 − Pq )W0 1 ·
b0,0 =¨ B ¸ ¨1 − (1 − Pq ) ¸ b0,0,e . (26)
¨ Pq ¸ ¨ Pq W0 ¸
© ¹ © ¹
−1
§ 1 − (1 − Pq )W0 · § 1 − (1 − Pq )W0 1 ·
Denote ¨
¨
B¸
¸ ¨¨1 − (1 − Pq ) ¸ in Eq. (26) by Ξ1 for
¸
© Pq ¹ © Pq W0 ¹
notational simplicity. Then substituting Eq. (26) into Eq. (21)-Eq. (25), b0, j ,e ,
18 M. Kim
j ∈ [1,W0 − 1], bi , j , i ∈ [1, m], j ∈ [−CW [q], Wi − 1], bm +1,0 can be expressed by b0,0,e .
Therefore, substituting these into the 2nd part of Eq. (20), we obtain b0,0,e as follows:
ª§ 1 · §1 (1 − Pq )2 − (1 − Pq ) 1 ·
b0,0,e = «¨ BΞ1 + (1 − Pq ) ¸ ⋅ ¨ (1 + PW − P ) + + W0 (W0 −1)(1 + C1Ξ1 ) ¸
W0 ¨ q 0 q 2 ¸
«©
¬ ¹ © Pq Pq 2 ¹
−1
°§ Aq +1 − Aqm++12 · § Aq +1 − Aqm++11 · °½ º
⋅ ®¨1 + ¸¸ + e(m) + d (m) + C (q) ⋅ ¨¨ 1 + ¸¸ ¾ ⋅ Ξ1 » , (27)
¨ 1 − Aq +1 1 − Aq +1 ¹ ¿° »¼
¯°© ¹ ©
Eq. (27) into Eq. (26), we obtain b0,0 and subsequently the stationary probabilities
{b0, j0 ,e , bi , j , bm +1,0 : i ∈ [0, m], j0 ∈ [0,W0 − 1] j ∈ [−CW [q],Wi − 1]} from Eq. (21)-Eq. (25).
With these stationary probabilities, we find the probability that a device in the priority
m
q class transmits during a unit backoff period is given by τ q = ¦ bi , − q .
i =0
4 Performance Analysis
4.1 Throughput
1−τ q
pI , q ∈ [0, Q ] ,
Let pB be the probability that the channel is sensed busy in a unit backoff period
which is given by pB = 1 − pI . Then pB − pS is the probability that the channel is
sensed busy by collisions that occur from any priority class. In addition, the
Performance Analysis of Service Differentiation for IEEE 802.15.4 Slotted CSMA/CA 19
probability that a collision occurs in a unit backoff period for the priority q class
which is denoted by pC , q is given by
Q
∏
nq −1
pC , q = 1 − (1 − τ q ) (1 − τ k ) nk , q ∈ [0, Q].
k = 0, k ≠ q
Let Sq be the normalized throughput for the priority q class. Then we can express
the normalized throughput Sq as the following ratio:
pS , qTpacket
Sq =
PI δ + pS TS + ( pB − pS )TC
Average delay and packet drop probability are other measures for the performance. In
this paper, however, we do not present them owing to the space limitation.
5 Numerical Results
In this section we present the performance of the analytical results, which show the
effect of the variations of CW[q] and BE[q] for service differentiation. We considered
BPSK mode and used the parameters in 802.15.4 standard, which are listed in
Table 1.
For the numerical result, we use all of the assumptions described in Sec. 2-3 such
as Poisson process of packet arrival and Q is taken by 2, in other words, 3 different
priority classes are considered. The arrival rate of each class is set by the same value
900 symbols during the time of unit backoff period which is 20/bpsk rate. Such an
arrival rate gives the probability of a new packet arrival by 0.5934.
For the service differentiation by BE, the value of CW[q] is set by 2 for all the
devices in every class and the values of BE[0], BE[1], and BE[2] at the device of each
class are set by 1, 2, and 3, respectively. On the other hand, for the service
differentiation by CW, we set BE[q] is set by 1 for all devices and the values of
CW[0], CW[1], and CW[2] are set by 2,4,8, respectively. In the following figures, x
axis denotes the number of devices of each class and we assume that the number of
devices in each class is all the same.
0.07 H 2 1 H 1
14000 14000 M 4 1 M 2
-2 L 3 0.05
0.06 L 8 10
12000 class CW class BE total throughput 12000 no priority 1 scale:1
H 2 H 1 CW
M 4 M 2 BE 0.05 0.04
10000 L 3 no priority 10000
L 8
no priority 0.04
8000 8000 0.03
0.03
6000 6000
0.02 0.02
4000 4000 0.01
0.01
2000 2000 0.00
0 0 -0.01 0.00
0 5 10 15 20 25 30 35 40 45 0 5 10 15 20 25 30 35 40 45
The number of devices in each class (N) The number of devices in each class (N)
The non-saturation throughputs for these values of CW[q] and BE[q] are shown in
Fig. 4. In addition, the throughput of no differentiation with CW=2 and BE=1 for all
devices and total throughputs of all of these are added. As see in the figure, for BE
differentiation, the throughput of a device on each class slightly increases as the total
number of devices increases and that of high class outperforms that of no priority
considered. On the other hand, for CW differentiation, it decreases for middles and
lower classes while that of high class increases regardless of the number of the
devices. It implies that the devices of lower classes yield the opportunity to occupy
the channel to the devices in higher class for CW differentiation. The delicate tuning
of the throughput could be performed by varying the value of BE[q] while the value
of CW[q] could be adjusted to increase the throughput of the high class. The total
throughputs of these cases have the following relation, except N = 2 , even though
the differences of performances are negligible: totalSCW < totalSno piority < totalSBE . For
N = 2 , we note that the total throughput of the case of no priority considered is
15,408 bits/sec while those of BE and CW differentiations are 11,596 bits/sec and
10,891 bits/sec, respectively. It dues to the fact that τ q and the success probabilities
out of τ q have the following relations:
where τ H , < and pS , H , < are the transmission probability and the probability of
successful transmission out of the transmission probability of high class in each
algorithm, respectively. In these relations, we note that for the higher class the more
chances of transmissions are given in CW differentiation, but the more successful
transmissions occur in BW differentiation.
Fig. 5 shows the probability of successful transmissions of three different service
classes. It shows that success probabilities increase as the number of devices increase
regardless of the priority classes for BE differentiation. On the other hand, for CW
differentiation, that of the high class is increasing while those of middle and low
Performance Analysis of Service Differentiation for IEEE 802.15.4 Slotted CSMA/CA 21
classes are decreasing as the number of devices is increasing. It implies the service
differentiation is greatly effected by CW differentiation rather than BE differentiation
just as same as throughput.. It is note worthy that the both probabilities of trying to
transmit and subsequent successful transmission with no priority are the highest
values compare to those values with differentiations are considered when N = 2 .
This is due by Eq. (29)-Eq. (30) and the fact that the probability of idle channel
without priority is lower than those with priority even though we do not present in this
paper. Actually, pI ,no priority = 0.7924 while pI ,CW = 0.9734 and pI , BE = 0.9684 , where
pI , < is given by Eq. (28) when < algorithm is applied. Regardless of the number of
devices, the relations τ H ,CW > τ H , BE , τ M ,CW < τ M , BE , and τ L ,CW > τ L , BE hold, where
H , M , and L are the high, middle, and low priority classes, respectively.
6 Conclusion
In this paper, we proposed two mechanisms for IEEE 802.15.4 sensor networks which
provide multiple level differentiated services for each and every device. The
mathematical model based on discrete-time Markov chain is provided for analyzing
the performance of the proposed mechanisms.
Numerical results show that the variation of contention window size has more
effect on service differentiation than that of backoff exponent as the number of
devices increase. The delicate tuning of the throughput could be performed by varying
backoff exponent while the better throughput of the high class could be performed by
adjusting contention window size.
For the future work, we will find out the optimal numbers of devices for different
packet arrival rates in the viewpoint of maximizing throughput and of minimizing
delay, which will provide a criterion for using the parameters for specific purposes.
References
[1] Standard for part 15.4: Wireless Medium Access Control (MAC) and Physical Layer
(PHY) Specifications for Low Rate Wireless personal Area Networks (WPAN), IEEE std.
802.15.4, IEEE, NY, 2003.
[2] J. Misic, S. Shafi, and V. B. Misic, The Impact of MAC Parameters on the Performance of
802.15.4 PAN, Elsevier Ad hoc Networks, vol. 2, pp. 351-371, July 2004.
[3] J. Misic, S. Shafi and V. B. Misic, Performance of a Beacon Enabled IEEE 802..15.4
Cluster with Downlink and Uplink Traffic, IEEE Transaction on Paralle; and Distributed
Systems, vol. 17, no. 4, pp 361-376, April 2006.
[4] T. R. Park, T. H. Kim, J. Y. Choi, S. Choi, and W. H. Kwon, Throughput and energy
consumption analysis of IEEE 802.15.4 slotted CSMA/CA, Eelectronics Letters, vol. 41,
no. 18, 1st Sep. 2005.
[5] T. H. Kim and S. Choi, Priority-based Delay Mitigation for Event-Monitoring IEEE
802.15.4 LR-WPANs, IEEE Communication Letters, vol. 10, issue 3, pp. 213-215, Mar.
2006.
22 M. Kim
[6] E.-J. Kim, M. Kim, S.-K. Youm, S. Choi, and C.-H. Kang, Priority-Based Service
Differentiation Scheme for IEEE 802.15.4 Sensor Networks, will be published in AEU-
International Journal of Electronics and Communications, 2006.
[7] J. Misic and V. B. Misic, Access delay for nodes with finite buffers in IEEE 802.15.4
beacon enabled PAN with uplink transmissions, Computer Communications, vol. 28,
pp1152-pp1166, 2005.
[8] B. Bougard, F. Catthoor, D. C. Daly, A. Chandrakasan, and W. Dehanene, Energy
Efficiency of the IEEE 802.15.4 Standard in Dense Wireless Microsensor Networks:
Modelling and Improvement Perspective, Design, Automation and Test in
Europe(DATE’05), vol. 1, pp.196-201, 2005.
Information-Driven Task Routing for Network
Management in Wireless Sensor Networks*
1 Introduction
Wireless sensors have the ability of sensing, computation and communication, and
work under the ad hoc mode to compose wireless sensor networks. Due to the small
volume of sensors, there are many resource constraints such as limited on-board
battery power and limited communication bandwidth, etc.. Because of the
characteristics in sensing ability and spatial coverage, WSN is ideally suited for
tracking moving targets, monitoring a large number of objects, detecting low-
observable events, and estimating the status of a target. The routing of these tasks is
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 23 – 32, 2006.
© Springer-Verlag Berlin Heidelberg 2006
24 Y. Liu et al.
not only to transfer data from one point to another, but should also be optimized for
the data transmission and the information aggregation.
This paper introduce an information-driven task routing algorithm to solve the
special routing scenario with communication cost constraints and sensor holes. It
utilizes the concept of information utility function to evaluate the expected mutual
information of the neighbor nodes, and selects the one that has the maximum
information contribution and satisfies the constraint of communication cost as the
next hop. And it calculates the information contribution of all the M hop neighbor
nodes to bypass the sensor holes.
Many algorithms were proposed to address the routing problem in WSNs. The
energy-aware routing [1] selected the path to minimize the energy exhaustion chance
of the sensors. Directed diffusion [2] routed data based on low-level data attributes
rather than sensor addresses. For the holes routing problem, the memorization based
algorithms [3] required nodes to record their past traffic or paths, and GPSR [4]
followed the perimeter of the hole to bypass it.
The idea of utilizing the information utility to manage sensing resources has been
investigated in computer fields already. [5] introduced an expected utility measure for
decentralized sensing system based on local decision. [6] proposed the information-
directed routing algorithm to track targets in WSNs, which illumines us to explore the
applications of the information-driven task routing (IDTR) algorithm for the routing
problems with communication cost constraints and sensor holes.
This paper is organized as follows. In section 2, the routing scenario of tracking the
target, the sequential Bayesian filtering, the information utility function and the two
related routing algorithms are introduced, and the IDTR-based heuristic search
algorithm is proposed. Section 3 is the simulation results and analyses.
For the application of target tracking in WSNs, the user may initiate a query from a
query proxy, requiring the sensor network to collect the target information and
estimate its location. The query proxy has to figure out where such information can be
collected and routes the query towards the high information content region, as
illustrated by the co-centric ellipses in Fig.1. This differs from the routing in
traditional communication networks where the destination is often known to the
proxy. Here, the destination is unknown and is dynamically determined. So the
geographical routing protocols are not suited.
Sometimes the aggregated information is required to be transferred to an exit node,
where it can be extracted for further processing, and the total communication cost is
also limited to be no more than a specified constraint [7], as illustrated in Fig.2.
Furthermore, the sparse distribution of sensors or the energy exhaustion of some
sensors may result in sensor holes. As illustrated in Fig.3, the two sensors in the
middle run out their deployed energy, resulting in a sensor hole as the dotted ellipse
area.
Information-Driven Task Routing for Network Management in WSNs 25
Exit
Fig. 1. Query to the high inf- Fig. 2. Query To Whom It Fig. 3. Sensor hole exists
ormation region May Concern: the exit node
For the tracking task in WSNs, we want to estimate the current position of the target
x ( ) based on the measurements of different sensors at different steps
t
{zi0
(0)
}
, zi1 (1) ," , zit ( t ) . IDTR utilizes sequential Bayesian filtering to estimate the
location of the target, evaluates the expected mutual information of the neighbor
nodes, and selects the one with maximum information contribution as the next hop.
measurement z (j
t +1)
to update the belief as follows:
( )
where p x ( ) | z ( ) is the prior belief given a history of the measurements up to time
t t
ª ( t +1
)
p x ( ) , zk( ) | z ( )
t +1 t
º
(
I MI ,k = MI X ( t +1) ( t +1)
; Zk |Z (t )
=z (t )
) = E (t+1) (t+1) (t ) log
« » (2)
p( x , zk |z ) «
¬ ( t +1 t
) (
p x ( ) | z ( ) p zk( ) | z ( )
t +1 t
) »
¼
For the scenario of routing the query from one proxy to the exit node with limited
communication costs, [7] proposed an A* heuristic search algorithm. It selected four
special paths as the representatives of the available paths that satisfied the
requirement of the total communication costs. The path with maximum accumulated
information gains was selected to perform sequential Bayesian filtering. The
computation complexity may be beyond the capability of the small volume sensors,
and the accumulated information gain can’t be computed by simply summing up the
values of several samples for its state dependency.
If sensor holes exist, [7] proposed a multiple step look-ahead algorithm. It
evaluated the information contribution of all the nodes in M hop neighborhood,
converted the graph representation of the network and tried to find the path with the
maximum accumulated information gain. The assumption is that the sum of
individual information ¦ I ( vk ) can often be considered as a reasonable
approximation of I ( v1 ,", vT ) in cases where the belief state varies slowly. Because
the neighbor sensors are close to each other, there is redundancy between I ( vk ) and
I ( vk −1 ) . So the assumption is not reasonable.
location. When selecting the next hop, the proposed algorithm not only considers the
information contributions of neighbors, but also the cost constraint C0 . The value of
C0 controls the tradeoff between the communication cost and the information
aggregation. Smaller C0 value favors shorter path, and larger C0 allows longer path
with more effective information aggregation.
Suppose that the path P ( t ) is planned up to the sensor vt , S ( ) is current belief at
t
(
Step 4. Forward the t + 1, S (
t +1)
) to v maxinfo , and the middle sensors on the shortest path
update the belief
28 Y. Liu et al.
( )
Once receiving the routing request t , S ( ) , sensor vt takes new measurement
t
( t +1) ( t +1)
z , and updates the belief to S through sequential Bayesian filtering. For
∀v j ∈ N vt , sensor vt calculates the sum of the distance, C j = v j − vt + v j − vexit .
C j < C means that v j is within the range of the ellipse, and its information
contribution needs to be calculated, assumed as I j . The one with the maximum
mutual information is selected as the next hop. Then construct a graph of vt ’s M hop
neighborhood, take the Euclidean distance as the weight of the edge between
neighboring sensors, and compute the shortest path from vt to vmaxinfo using
Dijkstra’s algorithm. The routing request t + 1, S ( ( t +1)
) is transferred to v maxinfo along
the shortest path. The middle sensors on the shortest path are only responsible for
update the belief and transfer the belief state. The vmaxinfo will repeat this belief update
and next hop selection process.
If there is no node within the range of the ellipse, select the sensor among N vt with
the smallest C j as the next hop. If the ellipse does not exist, the geographical routing
algorithm is used to search for the shortest path. The IDTR-based heuristic search
algorithm is summarized in Table.1.
Simulations are carried out to evaluate the performance of the proposed algorithm.
We use the grayscale grids to represent the mass distribution of the target location.
Blacker grid means that the target is more likely to be at the grid location. The initial
belief is assumed to be uniformly distributed over the entire sensor field.
Information-Driven Task Routing for Network Management in WSNs 29
100 acoustic amplitude sensors[8] are uniformly distributed in the 150 × 250 m 2
sensor field, and the range of radio is 30 m. The stationary target is located at
( 75,125) . We designate the sensor closest to the upper left corner as the query proxy
node, and the sensor closest to the lower left corner as the exit node.
Our algorithm is essentially a tradeoff problem between the communication cost
and the information aggregation controlled by the path length constraint C0 . We vary
the allowance C0 in {250,300, 350, 400, 450} . For each value, 20 independent
simulations are run to eliminate the effects of randomness. The statistical results are
listed in Table 2. With the increase of C0 , the average number of hops increases also,
while both the MSE and the belief size decreases. Compared with the shortest path
scenario, the IDTR-based heuristic search algorithm takes a little bit of detour, but
considerably improves the locating performance.
Fig.4 visualizes the selected paths and the grayscale graph of the target location under
different cost constraints. Fig.4(a) is the shortest path scenario. The path contains 13
hops and mostly follows a vertical line from the query proxy to the exit node. The
final location grayscale graph is a little bit dispersive. Fig.4(b) shows a longer path
with 15 hops. Starting from the query proxy, the path bends towards the target
direction and attempts to accumulate more information. The tracking performance is
vastly better than that of the shortest path scenario. Fig.4(c) shows a path with 17
hops. The tracking accuracy is further improved.
30 Y. Liu et al.
(a) MSE=7.0441, belief size=35 (b) MSE=4.8163, belief size=16 (c) MSE=2.9026, belief size=6
Fig. 4. Path selection and target location grayscale graph under different C0
We present simulation results for routing a query from an arbitrary query proxy node
to high information content region with sensor holes in the network. Sensor layout is
generated as follows: generate a uniform grid points of 15 rows and 6 columns to
evenly cover the region, and then perturb the grid points with independent Gaussian
noise of N ( 0,5) . To test the routing performance in the presence of sensor holes, we
remove the points in row 5 and 6 and columns 2-5, resulting a 4 × 2 routing hole. We
simulate the task of tracking moving target, which moves along the straight line
x = 75 with speed v = 7m / s .
Fig.5 shows several snapshots of the routing path applying our routing algorithm,
which represent the tracking scenarios at 20, 25 and 30 hops respectively. From the
figures we can see that in the IDTR-based heuristic search algorithm, because the
current sensor possesses the information of all M hop neighbors when selecting node,
the route can bypass the routing hole and extend to the target location, indicating that
the target locating is successful.
In Fig.5(a), even the target falls into the sensor holes, our algorithm still can bypass
the holes successfully, and collect information as more as possible around the target
to locate it. The true location is ( 75.0,106.0 ) , and the estimated position is
( 75.3,107.2 ) . In Fig.5(b), the path moving along with the target, occasionally reach
neighborhood to gather information. The true location is ( 75.0 141.0 ) and the
estimated position is ( 77.1 142.4 ) . In Fig.5(c), the true location is ( 75.0 176.0 ) and
the estimated position is. All realizes the object of correctly tracking.
The performance of our algorithm with the one proposed in [7] is compared. As
illustrated in Fig.6 (b), in order to arrive at the sensor with the maximum information
Information-Driven Task Routing for Network Management in WSNs 31
contribution, the path visits several middle sensors and combines their measurements
to refine the estimated belief. So the cloud is more compact. The numerical values are
listed in Table 3. Our algorithm achieves better target locating performance with the
cost of more communication hops.
(a) (b)
Fig. 6. Two snapshot of the multiple step look-ahead approach (a) and the IDTR-based heuri-
stic search method (b)
Number of hops 16 26
4 Conclusions
Tracking problem is one of the most important applications for WSNs. This paper
adopts information-driven task routing algorithm to solve the special routing scenario
with communication cost constraints and sensor holes. Sensors on the path update the
target belief via sequential Bayesian filtering, evaluate the information utility of M
hop neighbor sensors as well as the distance between the neighbors and the exit node,
and jointly optimize the selection of the next hop. The main goals of the proposed
algorithm are to minimize the communication cost and simultaneously to maximize
the accumulated information gain. In the future, we will apply the IDTR algorithm in
the tracking problem of several moving targets and the target estimation problem.
References
1. Shah, R.C., Rabaey, J.M.: Energy aware routing for low energy ad hoc sensor networks. In
Proc. IEEE Wireless Commun. Netw. Conf., Orlando, FL, (2001) 350-355
2. Intanagonwiwat, C. Govindan, R. Estrin, D. Heidemann, J. Silva, F.: Directed diffusion for
wireless sensor networking. IEEE/ACM Transactions on Networking. Volume 11, issue 1,
(2003) 2-16
3. Stojmenovic, I., Lin, X.: Loop-free hybrid single-path/flooding routing algorithms with
guaranteed delivery for wireless networks. IEEE Trans. Parallel Distrib. Syst., Volume 12,
No. 10, (2001) 1023-1032.
4. Karp, B., Kung, H.T.: Greedy perimeter stateless routing for wireless networks. In Proc.
MobiCom, Boston, MA, (2000) 243-254
5. Manyika, J., Durrant-Whyte, H.: Data Fusion and Sensor Management: A Decentralized
Information-Theoretic Approach. Ellis Horwood, New York (1994)
6. Feng, Zhao, Jaewon, Shin, Reich, J.: Information-driven dynamic sensor collaboration.
IEEE Signal Processing Magazine, Volume 19, Issue 2, (2002) 61-72
7. Liu, J., Feng Zhao, Petrovic, D.: Information-directed routing in ad hoc sensor networks.
IEEE Journal on Selected Areas in Communications, Volume 23, Issue 4, (2005) 851-861.
8. Juan. Liu, Reich, J.E., Feng, Zhao: Collaborative in-network processing for target tracking.
EURASIP, J. Appl. Signal Process, vol. 2003, Mar. (2003) 378-391
9. Haykin, S.: Adaptive Filter Theory. Prentice Hall Inc., 4th Ed., 2002
10. M., Chu, H., Haussecker, Feng, Zhao: Scalable information-driven sensor querying and
routing for ad hoc heterogeneous sensor networks. Int. J. High-Performance Comput.
Appl., vol. 16, no. 3. (2002) 293-313
11. T.M., Cover, J.A., Thomas: Elements of Information Theory. Wiley, New York, (1991)
Autonomic Management of Scalable Load-Balancing for
Ubiquitous Networks
1 Introduction
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 33 – 42, 2006.
© Springer-Verlag Berlin Heidelberg 2006
34 T. Tonouchi and Y. Beppu
In Section 2, previous load balancing methods are given. We show our algorithm
in Section 3. Section 3 also includes preliminary experiments, which shows a
problem of the proposed method. We clarify the mechanism causing the problem and
show how to prevent from the problem. Section 4 shows by experimental evaluations
the prevention succeeds.
Autonomic Management of Scalable Load-Balancing for Ubiquitous Networks 35
2 Related Work
A lot of valuable work has been done in the load balancing techniques. These works
are sorted into two types: one is a local load balancing technique and the other is a
global load balancing technique. Many vendors provide many kinds of local load
balancers. A local load balancer has a virtual IP address which is open to clients. The
load balancer dispatches incoming requests from the virtual IP address to several
back-end servers. It is easy to introduce local load balancers because an existing
server can be replaced with a load balancer and its back-end servers. In addition, the
system with a local load balancer can manage the back-end server failure because a
load balancer can easily monitor its back-end servers. However, a load balancer
itself can be bottleneck when huge traffics come. It is because a local load balancer
has only one virtual IP address and it must handle all traffics. As a result, local load
balancing technique has limitation in scalability.
A typical method of global load balancing techniques is a DNS round-robin
method; a special DNS server can be placed in the network (Fig. 2). The DNS server
answers the IP address of a not-busy server or a nearest server to a client which sends
a DNS request. The client, therefore, can access the not-busy server which is
expected to reply quickly or the nearest server which can communicate to the client
with a small network delay. However, caches of DNS answers result in the inaccurate
load balancing [3]. DNS servers consist of the tree structures whose roots are thirteen
root DNS servers in the world. A local DNS server may have caches of the answers
from its parent DNS server. The local DNS server may not know the change of the
load balancing situation because an “active” cache is used instead of the information
of its parent DNS server. To avoid cache problem, only one DNS server may gather
all the load information, and all the clients may access the DNS server. In this
method, however, the bottleneck of the DNS server emerges. It is said that the DNS
server can handle about thousands of requests per seconds. Because each server
access of a client accompanies with a DNS request, the DNS server is expected to
handles 1million – 10million [request/sec], but it may be impossible.
f5 3DNS system [4], which is now called Global Traffic Manager, tries to solve the
cache problem. 3DNS servers are allocated as local DNS servers, and they always
36 T. Tonouchi and Y. Beppu
㪠㫅㫋㪼㫉㫅㪼㫋
There are some activities in which Distributed hash table technique (DHT) is used
as a global load balancing mechanism. In short, jobs or contents are assigned to
servers in global networks by DHT. Jobs or contents are assigned the servers of IDs
same with the ID which a hash function applied to the jobs or contents calculated.
This may result in load balancing because the assignment by the hash function seems
to be random. Byes points out that this mechanism can not be a fair load balancing
[5]. Fig. 4 shows the problem when Chord [6] is used. Chord is not guaranteed to
assign servers into a “ring” in equal distances. A server whose ring distance between
itself and a “previous” server is long may be assigned more jobs.
Arc is short.
The number accesses is small.
3.1 Algorithm
Fig. 5 shows the intuitive image of the load exchange method. We assume that
servers know some servers (we call them neighbors), and this makes a server graph
G = (V, E) , where V is a set of servers, and E : (V, V ) is a set of a pair of a
server and its neighbor. We call E a set of edges. We assume that the number of
edges connected to a server is less than a constant number (e.g. 5) even if the number
of N becomes big. In our algorithm mentioned later, each server communicates with
its neighbors, and a large number of neighbors may become overhead.
“S.load” means the load of Server S ∈ V . A load may be a number of requests per
second, a CPU load, a response time, and so on. Server S and its neighbor S’ tell each
other their load information periodically. The number of neighbor servers is less than
given small constant. It means that the overhead and delay issued by this
communications are little because each server communicates with a small number of
neighbors.
Autonomic Management of Scalable Load-Balancing for Ubiquitous Networks 39
Sever S gives S’ some jobs when S.load / S’.load > D, where D is a given load
exchange threshold (D 1). In this case, S gives its job to S’ so that the load of S
becomes “S.load – l (S.load - S’.load) = (1-l) S.load + l S’.load” and so that the load
of S’ becomes “S’.load + l (S.load - S’.load) = (1-l) S’.load + l S.load”, where l is a
given load exchange factor (0 < l < 1). The load exchange is realized with redirection
mechanism. For example, in HTTP case, Server S issues 302 response with the URL
of Server S’. The client re-issues the request to S’ when it receives the response.
Server S repeats to answer a 302 response until the load becomes “(1-l) S.load +
S’.load”.
㪈㪇㪇
㪍㪇
Divergence
㪐㪇
㪌㪇
㪏㪇
load (call/sec)
load (call/sec)
㪎㪇
㪋㪇
㪍㪇
Convergence 㫊㫇㫀㪋
㫊㫇㫀㪌 㪊㪇
㪌㪇
㫊㫇㫀㪍
㫊㫇㫀㪎
㪋㪇
㪉㪇
㪊㪇
㪉㪇 㪈㪇
㪈㪇
㪇
㪈㪋㪑㪉㪇㪑㪇㪇
㪈㪋㪑㪉㪇㪑㪇㪌
㪈㪋㪑㪉㪇㪑㪈㪇
㪈㪋㪑㪉㪇㪑㪈㪌
㪈㪋㪑㪉㪇㪑㪉㪇
㪈㪋㪑㪉㪇㪑㪉㪌
㪈㪋㪑㪉㪇㪑㪊㪇
㪈㪋㪑㪉㪇㪑㪊㪌
㪈㪋㪑㪉㪇㪑㪋㪇
㪈㪋㪑㪉㪇㪑㪋㪌
㪈㪋㪑㪉㪇㪑㪌㪇
㪈㪋㪑㪉㪇㪑㪌㪌
㪈㪋㪑㪉㪈㪑㪇㪇
㪈㪋㪑㪉㪈㪑㪇㪌
㪈㪋㪑㪉㪈㪑㪈㪇
㪈㪋㪑㪉㪈㪑㪈㪌
㪈㪋㪑㪉㪈㪑㪉㪇
㪈㪋㪑㪉㪈㪑㪉㪌
㪈㪋㪑㪉㪈㪑㪊㪇
㪈㪋㪑㪉㪈㪑㪊㪌
㪈㪋㪑㪉㪈㪑㪋㪇
㪈㪋㪑㪉㪈㪑㪋㪌
㪈㪋㪑㪉㪈㪑㪌㪇
㪈㪋㪑㪉㪈㪑㪌㪌
㪈㪋㪑㪉㪉㪑㪇㪇
㪈㪋㪑㪉㪉㪑㪇㪌
㪈㪋㪑㪉㪉㪑㪈㪇
㪈㪋㪑㪉㪉㪑㪈㪌
㪈㪋㪑㪉㪉㪑㪉㪇
㪈㪋㪑㪉㪉㪑㪉㪌
㪈㪋㪑㪉㪉㪑㪊㪌
㪈㪋㪑㪉㪉㪑㪋㪇
㪈㪋㪑㪉㪉㪑㪋㪌
㪈㪋㪑㪉㪉㪑㪌㪇
㪈㪋㪑㪉㪉㪑㪌㪌
㪈㪋㪑㪉㪉㪑㪊㪇
㪇
㪈㪈 㪑㪊㪇
㪈㪈 㪑㪌㪋
㪈㪈 㪑㪇㪇
㪈㪈 㪑㪇㪍
㪈㪈 㪑㪈㪌
㪈㪈 㪑㪊㪇
㪈㪈 㪑㪊㪍
㪈㪈 㪑㪊㪐
㪈㪈 㪑㪋㪌
㪈㪈 㪑㪌㪈
㪈㪈 㪑㪊㪊
㪈㪈 㪑㪊㪍
㪈㪈 㪑㪊㪐
㪈㪈 㪑㪋㪉
㪈㪈 㪑㪋㪌
㪈㪈 㪑㪋㪏
㪈㪈 㪑㪌㪈
㪈㪈 㪑㪌㪎
㪈㪈 㪑㪇㪊
㪈㪈 㪑㪇㪐
㪈㪈 㪑㪈㪉
㪈㪈 㪑㪈㪏
㪈㪈 㪑㪉㪈
㪈㪈 㪑㪉㪋
㪈㪈 㪑㪉㪎
㪈㪈 㪑㪊㪊
㪈㪈 㪑㪋㪉
㪈㪈 㪑㪋㪏
㪈㪈 㪑㪌㪋
㪎
㪑㪌
㪎
㪎
㪑㪉㪍
㪑㪉㪍
㪍
㪑㪉㪍
㪍
㪍
㪑㪉㪍
㪍
㪑㪉㪍
㪍
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉㪎
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪑㪉
㪈㪈
Time
Time
Conditions: Complete graph of four nodes,
Conditions: Complete graph of four nodes,
D = 1.2, l = 0.8
D = 2, l = 0.6
load load
load
load load
load
load
Fig. 7 shows the reason of the divergence. When a server with a light load (at the
center of the figure) is connected to servers with a heavy load (step 1), the server with
light load receives the load from its connected servers. As a result, it must manage
heavy load (Step 2). It, then, gives its load into another connected server (e.g. a
server in right side) (Step 3). And the right side server will try to give its load into
another server in the next time (Step 4).
xt = (x0,t " x n−1,t ) where t means a transpose matrix. The load exchange is
G def t
modeled by :
xi ,t +1 = xi ,t − l ¦ (x i ,t − x j ,t ) (1)
j∈N ( i )
¦a i, j =0 (2)
j =0
because the number of neighbors equals to the sum of edges. We can express (1) as
G G
xt +1 = (−lA + E ) xi
G G (3)
xt = (−lA + E ) t −1 x0
where E is a unit matrix.
Autonomic Management of Scalable Load-Balancing for Ubiquitous Networks 41
def
It is known that lim xi ,t converge if all eigen values of H = (−lA + E) are
t →∞
larger than -1, and 1 or below. It is clear that det(H − αE) = 0 where Eigen value
α H . One of the eigen values of G is 1 because
of
det(H − αE) = det(H − E) = det(−lA) = l n det( A) if α = 1 . And then, you
can easily find det( A ) = 0 because of (2). In addition, the eigen vector
G t G G
corresponding to Engen value α = 1 is α i = (1"1) because Hα i = α i . This eigen
vector means that all loads are equalized in future. In short, the sufficient condition
of the convergence is that all the eigen values α of H are − 1 < α ≤ 1 . All loads
converge in the equal loads if the sufficient condition is satisfied.
4 Experimental Evaluations
In order to prove the correctness of the mathematical model, we compare the
simulation results based on the mathematical model with the experimental
evaluations. We give burst requests to a server, and Fig. 8 shows the loads of servers
after the burst requests. The upper three graphs are the results of the simulations and
the lower three graphs are those of the experimental results. Each pair of graphs in
the same row is evaluated under the same condition. The left pair is evaluated under
l = 0.4 , the center pair is evaluated under l = 0.6 , and the right pair is evaluated
under l = 0.8 . Four servers are connected to the other servers in all evaluations. In
short, servers in each evaluation make a complete graph.
Both graphs in the left pair converge because the max absolute value of eigen
values except 1 is -0.84 in the left pair. Both of the graphs in the right pair diverge
because the max absolute value of eigen values is -2.7 in the right pair. These two
䊉䊷䊄᭴ᚑ㩷㩿㫅㪔㪋㪀㩷ቢో䉫䊤䊐
complete graphs with four nodes 䊉䊷䊄᭴ᚑ㩷㩿㫅㪔㪋㪀㩷ቢో䉫䊤䊐
complete graphs with four nodes 䊉䊷䊄᭴ᚑ㩷㩿㫅㪔㪋㪀㩷ቢో䉫䊤䊐
complete graphs with four nodes
(simulation) (simulation) (simulation)
㪈㪅㪉 㪊㪇 㪊㪇㪇㪇
㪈
Convergence 㪉㪇
Divergence 㪉㪇㪇㪇
load (call/sec)
㪇㪅㪏 㪈㪇㪇㪇
㪈㪇
load (call/sec)
㪇㪅㪍 㪇
⽶⩄
䊉䊷䊄㪈 㪇
⽶⩄
䊉䊷䊄㪈
⽶⩄
㪇㪅㪋 䊉䊷䊄㪈 㪇 㪈 㪉 㪊 㪋 㪌 㪍 㪎 㪏 㪐 㪈㪇 㪈㪈
㪄㪈㪇㪇㪇
䊉䊷䊄㪉 㪄㪈㪇 㪇 㪈 㪉 㪊 㪋 㪌 㪍 㪎 㪏 㪐 㪈㪇 㪈㪈 䊉䊷䊄㪉 䊉䊷䊄㪉
㪇㪅㪉 䊉䊷䊄㪊 㪄㪉㪇㪇㪇
䊉䊷䊄㪊 䊉䊷䊄㪊
㪇 㪄㪉㪇
Divergence
㪄㪊㪇㪇㪇
㪄㪇㪅㪉 㪇 㪈 㪉 㪊 㪋 㪌 㪍 㪎 㪏 㪐 㪈㪇 㪈㪈 㪄㪊㪇 㪄㪋㪇㪇㪇
㪄㪇㪅㪋 㪄㪋㪇 㪄㪌㪇㪇㪇
ᤨ㑆 Time Time
ᤨ㑆(tic)
Time (tic) ᤨ㑆(tic)
Evaluation
㪌㪇 㪌㪇
Evaluation 㪌㪇
Divergence Evaluation
㪋㪇 㪋㪇 㪋㪇
㪊㪇
Convergence 㫊㫇㫀㪋
㫊㫇㫀㪌
㫊㫇㫀㪍
㪊㪇
Convergence 㫊㫇㫀㪋
㫊㫇㫀㪌
㫊㫇㫀㪍
㪊㪇
㫊㫇㫀㪋
㫊㫇㫀㪌
㫊㫇㫀㪍
㫊㫇㫀㪎
㫊㫇㫀㪎 㫊㫇㫀㪎
㪉㪇
㪉㪇 㪉㪇
㪈㪇
㪈㪇 㪈㪇
㪇
㪈㪋㪑㪇㪋㪑㪇㪇
㪈㪋㪑㪇㪋㪑㪈㪉
㪈㪋㪑㪇㪋㪑㪈㪏
㪈㪋㪑㪇㪋㪑㪉㪋
㪈㪋㪑㪇㪋㪑㪊㪇
㪈㪋㪑㪇㪋㪑㪋㪉
㪈㪋㪑㪇㪋㪑㪋㪏
㪈㪋㪑㪇㪌㪑㪇㪇
㪈㪋㪑㪇㪌㪑㪈㪉
㪈㪋㪑㪇㪌㪑㪈㪏
㪈㪋㪑㪇㪌㪑㪊㪇
㪈㪋㪑㪇㪌㪑㪋㪉
㪈㪋㪑㪇㪌㪑㪋㪏
㪈㪋㪑㪇㪌㪑㪌㪋
㪈㪋㪑㪇㪍㪑㪇㪇
㪈㪋㪑㪇㪍㪑㪈㪉
㪈㪋㪑㪇㪍㪑㪉㪋
㪈㪋㪑㪇㪍㪑㪊㪇
㪈㪋㪑㪇㪍㪑㪋㪉
㪈㪋㪑㪇㪍㪑㪋㪏
㪈㪋㪑㪇㪎㪑㪇㪇
㪈㪋㪑㪇㪎㪑㪈㪉
㪈㪋㪑㪇㪎㪑㪉㪋
㪈㪋㪑㪇㪎㪑㪊㪇
㪈㪋㪑㪇㪎㪑㪋㪉
㪈㪋㪑㪇㪎㪑㪌㪋
㪇
㪈㪋㪑㪇㪋㪑㪇㪍
㪈㪋㪑㪇㪋㪑㪊㪍
㪈㪋㪑㪇㪋㪑㪌㪋
㪈㪋㪑㪇㪌㪑㪇㪍
㪈㪋㪑㪇㪌㪑㪉㪋
㪈㪋㪑㪇㪌㪑㪊㪍
㪈㪋㪑㪇㪍㪑㪇㪍
㪈㪋㪑㪇㪍㪑㪈㪏
㪈㪋㪑㪇㪍㪑㪊㪍
㪈㪋㪑㪇㪍㪑㪌㪋
㪈㪋㪑㪇㪎㪑㪇㪍
㪈㪋㪑㪇㪎㪑㪈㪏
㪈㪋㪑㪇㪎㪑㪊㪍
㪈㪋㪑㪇㪎㪑㪋㪏
㪇
㪈㪉 㪑㪈 㪏
㪈 㪉 㪑㪉 㪋
㪈㪈 㪑㪇㪇
㪈 㪈 㪑㪇 㪊
㪈㪈 㪑㪇㪍
㪈 㪈 㪑㪇 㪐
㪈㪈 㪑㪈㪉
㪈 㪈 㪑㪈㪌
㪈㪈 㪑㪈㪏
㪈㪈 㪑㪉㪈
㪈㪈 㪑㪉 㪋
㪈㪈 㪑㪉㪎
㪈 㪈 㪑㪊 㪇
㪈㪈 㪑㪊㪊
㪈 㪈 㪑㪊 㪍
㪈㪈 㪑㪊㪐
㪈 㪈 㪑㪋㪉
㪈㪈 㪑㪋㪌
㪈㪈 㪑㪋㪏
㪈㪈 㪑㪌 㪈
㪈㪈 㪑㪌㪋
㪈 㪉 㪑㪌 㪎
㪈㪉 㪑㪇㪇
㪈 㪉 㪑㪇 㪊
㪈㪉 㪑㪇㪍
㪈 㪉 㪑㪇㪐
㪈㪉 㪑㪈㪉
㪈 㪉 㪑㪈㪌
㪈 㪉 㪑㪉 㪈
㪑㪉㪎
㪈㪉 㪑㪊㪇
㪈㪉 㪑㪊㪋
㪈㪉 㪑㪊㪍
㪈㪉 㪑㪌㪇
㪈㪉 㪑㪌㪍
㪈㪉 㪑㪇㪏
㪈㪉 㪑㪈㪇
㪈㪉 㪑㪉㪏
㪑㪊㪇
㪈㪉 㪑㪊 㪉
㪈㪉 㪑㪊㪏
㪈㪉 㪑㪋㪇
㪈㪉 㪑㪋㪉
㪈㪉 㪑㪋㪋
㪈㪉 㪑㪋 㪍
㪈㪉 㪑㪋 㪏
㪈㪉 㪑㪌 㪉
㪈㪉 㪑㪌㪋
㪈㪉 㪑㪌㪏
㪈㪉 㪑㪇㪇
㪈㪉 㪑㪇 㪉
㪈㪉 㪑㪇 㪋
㪈㪉 㪑㪇 㪍
㪈㪉 㪑㪈㪉
㪈㪉 㪑㪈㪋
㪈㪉 㪑㪈㪍
㪈㪉 㪑㪈㪏
㪈㪉 㪑㪉 㪇
㪈㪉 㪑㪉 㪉
㪈㪉 㪑㪉 㪋
㪈㪉 㪑㪉 㪍
㪑㪇 㪇
㪑㪌㪐
㪐
㪑㪌㪐
㪐
㪑㪌㪐
㪑㪌㪐
㪑㪌㪐
㪑㪌㪐
㪑㪌 㪐
㪑㪌㪐
㪐
㪑㪌㪐
㪐
㪑㪌㪐
㪑㪌㪐
㪑㪌㪐
㪑㪌㪐
㪑㪌 㪐
㪑㪌㪐
㪐
㪑㪇㪇
㪑㪇 㪇
㪑㪇㪇
㪑㪇㪇
㪑㪇㪇
㪑㪇 㪇
㪑㪇 㪇
㪑㪇㪇
㪑㪇
㪑㪌
㪑㪌
㪑㪌
㪑㪌
㪑㪌
㪑㪈㪇
㪑㪈㪇
㪑㪈㪇
㪇
㪑㪈㪇
㪇
㪑㪈㪇
㪑㪈㪇
㪇
㪑㪈㪇
㪇
㪈
㪑㪈㪈
㪑㪈㪈
㪑㪈㪈
㪑㪈㪈
㪈
㪈
㪑㪈㪈
㪈
㪑㪈㪈
㪑㪈㪈
㪑㪈㪈
㪑㪈㪇
㪑㪈㪇
㪑㪈㪇
㪑㪈㪇
㪑㪈㪈
㪑㪈㪈
㪑㪈㪈
㪑㪈㪈
㪈㪈
㪑㪈
㪑㪈
㪑㪈
㪑㪈
㪑㪈
㪑㪈
㪑㪈
㪑㪈
㪈㪉
results show that the model is correct in the experiments. Notice that the lower graph
in the center pair converges while the upper one diverges. We think that the
convergence condition given in Section 0 is sufficient condition. Therefore, there
some cases where the system converges while the corresponding model divergences.
5 Conclusion
We propose the naïve global load balancing method in this paper. The method is
prone to diverge, but we model the method and clarify the sufficient condition that the
load balancing converges. The correctness of the model is proved through the
experimental result.
This algorithm assumes that Graph G is given. We are now developing the
autonomic graph making protocols. We aim at the maintenance-free and fault-tolerant
graph construction.
Acknowledgement
This research was sponsored by the Ministry of internal Affairs and Communication
in Japan.
References
1. F5, Broadband-Testing: Application Traffic Management , Jan., 2005
2. T. Berners-Lee, et al., Hypertext Transfer Protocol -- HTTP/1.0, RFC 1945, May 1996
3. Tony Bourke, Server Load Balancing, O’reilly, 2001, ISBN 0-596-00050-2
4. f5, “BIG-IP Global Traffic Manager”, http://www.f5.com/products/bigip/gtm/
5. Byes, J., et al. Simple Load Balancing for Distributed Hash Tables, in Proceedings of 2nd
International Workshop on Peer-to-Peer Systems (IPTPS '03), pp. 80-87
6. Stoica, I., et al. Chord: A scalable peer-to-peer lookup service for internet applications, In
ASM SIGCOMM 2001, pp.149-160
7. The Apache Software Foundation, Apache Tomcat, http://tomcat.apache.org/
A Policy-Based Management Framework for
Self-managed Wireless Sensor Networks
1 Introduction
In Wireless sensor networks (WSNs), radio bandwidth is scarce, computational
power is limited, and energy efficient is paramount. Such limitations are chal-
lenges to overcome. In particular, one of the essential needs is for a system that
autonomously manages the limited energy and bandwidth of WSNs. In WSNs,
a number of sensor nodes are deployed over a large area and long distances and
multi-hop communication is required between nodes and sensor nodes have the
physical restrictions in particular energy and bandwidth restrictions. So man-
aging numerous wireless sensor nodes directly is very complex and is not effi-
cient. To make sensor nodes perform intelligent self-management, they should
be organized and managed automatically and dynamic adjustments need to be
done to handle changes in the environment. In [1], we propose an autonomous
The present research has been conducted by the Research Grant of Kwangwoon
University in 2006.
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 43–52, 2006.
c Springer-Verlag Berlin Heidelberg 2006
44 J.-E. Lee et al.
2 Backgrounds
2.1 Management Architectures
Linnyer B. Ruiz designed the MANNA architecture [4] for WSNs, which con-
siders three management dimensions: functional areas, management levels, and
WSN functionalities. He also proposed WSN models to guide the management
activities and the use of correlation in the WSN management. However, he de-
scribed only conceptual view of the distribution of management functionalities
in the network among manager and agent.
Mohamed Younis [5] proposed architecture for monitoring and management
of sensor networks, which focuses on reducing the sensitivity of the operation
and monitoring of sensor networks to the ambiguity of the propagation model
of the radio signal. He suggested agent sensors that relay messages to and from
unreachable sensors and groups of sensors around these agents while considering
the load on each agent.
Chien-An Lee [6] proposed an intelligent self-organization management mech-
anism for sensor networks. The nodes are classified into three levels according
to their functionality. The nodes in the low level are managed by those in the
higher level and form hierarchical management structures. His work indicates
how high-level nodes form a cluster through a contest with low-level nodes. Per-
formance measures the cover loss, the average delay between the header and the
member nodes and the 20∼80 rules are also considered.
wt w Gt yGt
h w Gk
xz t
w Gt l Gt
p t
nGw
yGt
s¡Gz { Gt
G
OwhP
OP
{GY
OP
{GX
OGP
w
l
sGk
kGm wl
{GW
OGP
sGw
and/or QoS management go through an aggregation and fusion phase when en-
ergy and/or QoS information collected are merged and fused into energy and/or
QoS contours by means of cluster heads.
The PE enforces local policies assigned by the PM to make local decisions
and filter off unessential redundant sensed data. To do this, the PE consists of
policy enforcement function, local decision function, data filtering function, and
local policies. The PE communicates with the PA via TinyCOPS-PR protocol
to be assigned local policies.
compares its node ID with the region ID (r id) from the messages. If the previous
comparison is true, node k declares itself as a region node (r node) and sets its
region ID to its node ID. Otherwise, if node k ’s node ID is included in a special
region list (r list) from the message, node k sets its region ID to a corresponding
region ID of the message. The region node selection is completed with region
confirmation messages (r conf msgs) broadcasted from all of cluster heads.
ͣͦ͡
ͣ͡͡
ͶΟΖΣΘΪ͑ΔΠΟΤΦΞΡΥΚΠΟ͙͑ͻ
ͦ͢͡
ͽͶͲʹ
ͽͶͲʹ͞ʹ
΄ͿΈʹͽΆ΄΅Ͷ
͢͡͡
ͦ͡
͡
͢ ͣ ͤ ͥ ͦ ͧ ͨ ͩ ͪ ͢͡
ͿΦΞΓΖΣ͑ΠΗ͑ΔΝΦΤΥΖΣΤ͑ΘΖΟΖΣΒΥΖΕ
In the case of LEACH clustering algorithm, until the number of clusters gen-
erated is 2, it shows significantly higher energy consumption compared to the
other clustering algorithms, but after generation of more than 3, it is stabilized
showing gradual increase. LEACH-C Clustering Algorithm shows progressive in-
crease in energy consumption from round 1 to round 10. Similar to LEACH-C,
SNOWCLUSTER algorithm also shows results of gradual increase, but its con-
sumption rate is slightly less than that of LEACH-C.
Fig. 3 is the result showing the amount of energy that is consumed during
transmission of management message to the sensor node from the base station
after formation of three clusters in the network topology of 200 nodes.
A Policy-Based Management Framework for Self-managed WSNs 49
In the case of LEACH, because it does not have the position information
of the nodes, inefficient routing is being resulted, and as a result, significantly
greater amount of energy is consumed in transmitting management messages.
ͶΟΖΣΘΪ͑ΔΠΟΤΦΞΡΥΚΠΟ͙͑ͻ ͨ
͡
ͽͶͲʹ ͽͶͲʹ͞ʹ ΄ͿΈʹͽΆ΄΅Ͷ
5 Implementation
5.2 Snowman
The PM and the PA of SNOWMAN architecture are implemented on Windows
XP systems using pure JAVA. The PE is implemented on TinyOS in the Nano-24
nodes using gcc.
Fig. 5 shows the input forms for policy information on the PM. We use the
XML technologies to define and handle global policies. There are several advan-
tages of using XML in representing global policies [11]. Because XML offers many
useful parsers and validators, the efforts needed for developing a policy-based
management system can be reduced. To define XML policies, we customized and
used the Scott’s XML Editor [12]. The defined policies are stored locally in the
policy storage of the PM and are stored remotely in the policy storage of the
PA. PM communicates with PA via simple ftp for policy transmissions. To policy
distribution to sensor nodes, we also design and implement TinyCOPS-PR that
is simplified suitably for wireless sensor networks.
A Policy-Based Management Framework for Self-managed WSNs 51
6 Conclusion
In this paper, we proposed a policy-based management framework for self-
managed WSNs called SNOWMAN. This paper also presented SNOWCLUSTER
clustering algorithm. The SNOWMAN framework integrated the policy-based
management paradigm and hierarchical cluster-based management architecture.
SNOWMAN can provide administrators with a solution to simplify and auto-
mate the management of WSNs using PBNM paradigm. SNOWMAN can also
reduce the costs of managing sensor nodes and of the communication among
them using hierarchical clustering architecture. SNOWCLUSTER algorithm can
be said to be more efficient in the aspect of network management and energy
consumption than other existing sensor network clustering algorithms.
References
1. Si-Ho Cha, et al., A Self-Management Framework for Wireless Sensor Networks,
LNCS 3842, January 2006.
2. Kaustubh S. Phanse, Luiz A. DaSilva, Extending Policy-Based Management to
Wireless Ad Hoc Networks, 2003 IREAN Research Workshop, April 2003.
3. R. Yavatkar, D. Pendarakis, R. Guerin, A Framework for Policy-based Admission
COntrol, IETF RFC 2753, January 2000.
4. Linnyer B. Ruiz, Fabircio A. Silva, Thais R. M. Braga, José M. Nogueira, An-
tonio A. F. Loureiro, On Impact of Management in Wireless Sensors Networks,
IEEE/IFIP NOMS 2004, Volume 1, 19-23 April 2004.
5. M. Younis, P. Munshi, Architecture for Efficient Monitoring and Management of
Sensor Networks, IFIP/IEEE E2EMON, September 2003.
6. Chien-An Lee et al., Intelligent Self-Organization Management Mechanism for
Wireless Sensor Networks, http://www.ndhu.edu.tw/ rdoffice/exchange/CYC-
paper.pdf.
7. M. J. Handy, M. Haase, D. Timmermann, Low Energy Adaptive Clustering Hier-
archy with Deterministic Cluster-Head Selection, 2002 IEEE.
52 J.-E. Lee et al.
1 2 3
Atsushi Kobayashi , Daisuke Matsubara , Shingo Kimura ,
3 1 1
Motoyuki Saitou , Yutaka Hirokawa , Hitoaki Sakamoto ,
1 1
Keisuke Ishibashi , and Kimihiro Yamamoto
1
NTT Information Sharing Platform Laboratories,
3-9-11 Midori-cho, Musashino, Tokyo, 180-8585 Japan
2
Hitachi, Ltd., Central Research Laboratory,
1-280 Higashi-Koigakubo, Kokubunji, Tokyo, 185-8601 Japan
3
NTT Advanced Technology Corporation,
3-9-11 Midori-cho, Musashino, Tokyo, 180-8585 Japan
1
[email protected], {hirokawa.yutaka, sakamoto.hitoaki,
ishibashi.keisuke, yamamoto.kimihiro}@lab.ntt.co.jp,
2 3
[email protected], [email protected],
[email protected]
1 Introduction
Recently, there has been growing interest in flow-based traffic measurement methods,
such as sFlow [1] and NetFlow [2], because these methods are useful for anomaly-
detection and traffic engineering. These methods are based on traffic information
exporting from routers.
The other sides, the traffic volumes handled by ordinary service provider networks
have increased year-by-year. In Japan, the number of broadband users increased
significantly and broadband traffic consumes 230Gbps. In next five years, traffic
volume passing through the point of presence of major service providers is likely to
reach 100Gbps. In such a traffic volume, we need to reduce traffic records to reduce
the burden on the traffic monitoring system and handle properly them. Therefore,
generally sampling and aggregation method are applied in router. But, these flow
records reduced by sampling until required accuracy of anomaly detection are still too
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 53 – 62, 2006.
© Springer-Verlag Berlin Heidelberg 2006
54 A. Kobayashi et al.
large for monitoring system to handle them. In aggregation method, generally flow
records are aggregated based on network prefix or AS number. But, we can not
monitor detailed traffic information. In such established practices, there is the issue
that we can’t monitor detailed information over handling large traffic volume. Thus,
we need a method that provides efficient aggregation and storage function, and need
the architecture of allowing aggregation by multi-granularity.
We focus on flow concentrator that is one of IPFIX nodes in the [6]. There is a
possibility that it achieves an aggregation by multi-granularity, however the detailed
internal process model is not sufficiently described yet in other documents. We
propose the model of flow concentrator that has aggregation, storage and distribution
functions. The main contribution of this paper is the proposal of architecture for a
traffic monitoring system using flow concentrators. Using flow concentrators enables
achievement for large-scale traffic monitoring system.
The organization of this paper is as follows. An introduction to flow concentrators
is given in Section 2. Section 3 proposes internal and external reference models of
flow concentrators that achieve a large-scale traffic monitoring system. Section 4
discusses implementation issues of flow concentrators. Section 5 describes the flow
concentrator called “FlowSquare” that we developed and presents evaluation results.
O: Observation point
M: Metering process
E: Exporting process
C: Collecting process
− It provides functions for balancing the aggregation load and storing aggregated
data without requiring the receiving traffic collectors to have any special function.
− It has dual-role architecture for receiving and sending flow records. This
architecture allows cascading concentrators, and the number of flow concentrators
can be adjusted to suit the size of a given network.
− By storing flow records before aggregation, it enables us to refer to detailed traffic
information in addition to summary information. This increases the scalability.
− External nodes can look up distributed flow records that have been stored in each
flow concentrator.
Our reference model for a flow concentrator based on the above requirements is
presented in the next section.
NetFlow v9 NetFlow v9
IPFIX IPFIX Traffic
Router
collector
Flow
concentrator
Flow Flow
concentrator concentrator
To coordinate with flexible flow transport protocols like NetFlow v9 or IPFIX, the
internal process model needs a flexible model. We propose a combination of several
components as the internal process model. In Fig. 3, the flow concentrator contains
six components: the collecting, selection, aggregation, reporting, exporting and
storing processes. The collecting process receives flow records from routers. It also
forwards received flow records to multiple selection processes or storing processes.
The selection process has a filtering function and selects flow records that are
matched under given conditions. The storing process has selects specified information
elements by using storage rules and stores these flow records in a database. The
aggregation process creates aggregated flow records in accordance with aggregation
rules that are described in the aggregation draft [7]. The reporting process manages
the reporting template, and the exporting process forwards flow records to the next
nodes. Several processes are described in reference model document [8].
56 A. Kobayashi et al.
Flow concentrator
Flow
records Collecting Selection Aggregation Reporting
process process process process
Through
NetFlow v9
IPFIX
Flow
Exporting records
Storing Data
process process
base Through
NetFlow v9
IPFIX
Traffic
collector
IPFIX
Second-level NetFlow v.9
Packet
Packet router router
router
router router router
4 Implementation Issues
Each process performs functions based on instruction parameters that are set by an
operator in advance. Some of them should be set by information elements and,
defining instruction words enables the creation of the instruction templates, just like
flow template of NetFlow v9 or IPFIX.
For example, the aggregation process has instruction words like "key", "keep", and
"discard". This increases the flexibility, as in [7]. Similar to aggregation process, the
storing process can have storing instruction templates. If the instruction word "store"
is set in some information elements, these information elements should be stored in
database. If "discard" is set as the instruction word, these information elements should
not be stored. Examples of the storing and the aggregating instruction templates are
shown in Fig. 6. The storing process stores only information elements that are labeled
with the instruction "store" in the database. The aggregation process creates
aggregated flow records that have common source/destination network addresses
because the "key" instruction word is set for the "Source IP address", "Destination IP
address", "Source address prefix", and "Destination address prefix" elements.
Additionally, we have proposed using those instruction templates as a management
information base (MIB) [11]. External nodes need to be able to control several flow
concentrators through remote access in order to control the whole system. This
58 A. Kobayashi et al.
Fig. 6. Examples of the storing instruction template and aggregation instruction templates
In general, traffic collector uses an RDBMS that has an SQL query function.
However, once flow records have been stored in the RDBMS, they are not changed
within a certain storage interval. Thus, the RDBMS approach cannot be said to be the
optimal solution for a large scale and high speed. Many recent open source tools for
use as collectors have a flat file structure [12, 13]. This section discusses these
solutions and considers which is suitable for a traffic collector that has several
applications or for a flow concentrator that needs scalability.
Since PostgreSQL and MySQL have been widely used as RDBMSs and we
focused on them in our evaluation. MySQL offers several engine types. We
considered two of them MyISAM and HEAP. Of these, HEAP is fast because it is a
memory-resident function.
As a flat file structure, we considered a method based on Perl storable which is
useful for quick development and a flat file structure. The flat file structure obtained
with Perl storable is outline in Fig. 7. With respect to directory structure, the root
directory is set to the exporter address which is the router’s address. The next
directory is set to the date which means the year, month and day. The name of each
file is its time of storage. This is useful when an operator is searching flow records for
a specific time and deleting old flow records. With respect to storable structure, one
flow record is converted into one hash table in order to make elements easy to search.
The references of each hash table are added to the array table. Our test tool saves
storable files at intervals of 3 or 5 minutes.
We examine time taken for storing, searching, and deleting each solution in the
following environment. The searching time was included in the time for displaying
A Proposal of Large-Scale Traffic Monitoring System Using Flow Concentrators 59
for searched flow records. Additionally, we evaluated the influence of the RDBMS
when the number of accumulation records was increased.
− CPU: Intel(R) Pentium(R) 4 3.40GHz
− Memory :3574MB
− OS: FreeBSD 5.4
− Perl version 5.8.6
− MySQL version 4.1.16
− PostgreSQL version 8.0.7
Figure 8 shows the time required for storing and searching flows for each solution.
The deleting time was about one second in all solutions and for all flow records. This
indicates that the flat file structure using "storable" is preferable from the viewpoint of
speed. Generally, the performance of an RDBMS is likely to become slower as the
accumulation volume increases. Figure 9 shows the influence of RDBMS versus
accumulation volume. In particular, the searching and deleting times clearly became
slower as the number of accumulation records increased. In a large network, about
100 million records are accumulated per day. Therefore, it might be difficult to use an
RDBMS.
Fig. 8. Time required for storing (a) and searching (b). For storage, a flat file structure using
Perl storable is faster than other methods. On the other hand, Perl storable is slow at searching
and displaying, but this is less critical.
60 A. Kobayashi et al.
Table 1 summarizes the results for the solutions, including streaming database,
which has been proposed as a high-speed query function [14]. We can select these
solutions case by case. For example, a flat file structure is preferable when the node
needs scalability, just like a flow concentrator. On the other hand, RDBMS or a
memory-resident database is preferable when the node needs multiple functions, just
like a traffic analyzer.
Fig. 10. Test environment. The traffic matrix measurements system was located in the traffic
analyzer. We examined the CPU load on the analyzer and concentrator.
Fig. 11. Load on traffic analyzer in both cases. The load on the analyzer was reduced when a
concentrator was used.
Fig. 12. CPU load of flow concentrator and traffic analyzer versus exporting speed of records.
Also shown is the change in aggregation rate with record exporting speed.
In particular, the flow concentrator is useful for a top-level collector that needs all
traffic information, like a traffic matrix measurement system. We can get more detailed
flow records from flow concentrators and can search for more detailed information that is
correlated to the whole traffic summary. This is, we can drill down.
6 Conclusion
In this paper, we proposed a reference model for a large scale traffic monitoring
system using flow concentrators. These models include an external model as the
solution and an internal process model to achieve greater scalability. In addition, we
developed a flow concentrator that has basic functions and examined the effectiveness
of using flow concentrators. In future works, we will develop multiple functions in
flow concentrators to further increase the scalability and flexibility.
Acknowledgments
This study was supported by the Ministry of Internal Affairs and Communications of
Japan.
References
1. P. Phaal, S. Panchen and N. McKee, “InMon Corporation’s sFlow: A Method for
Monitoring Traffic in Switched and Routed Networks,” RFC3176.
2. http://www.cisco.com/en/US/products/ps6601/products_ios_protocol_group_home.html
3. B. Claise, “Cisco Systems NetFlow Services Export Version 9," RFC3954.
4. B. Claise, “IPFIX Protocol Specification,” draft-ietf-ipfix-protocol-16.txt(work in
progress)
5. B. Claise, “Packet Sampling (PSAMP) Protocol Specifications,” draft-ietf-psamp-
protocol-02.txt (work in progress).
6. J. Quittek, T. Zseby, B. Claise, and S. Zander, “Requirements for IP Flow Information
Export(IPFIX),” RFC3917.
7. F. Dressler, C. Sommer, and G. Munz, "IPFIX Aggregation," draft-dressler-ipfix-
aggregation-01.txt (work in progress).
8. A. Kobayashi, K. Ishibashi, K. Yamamoto and D. Matsubara, “The reference model of
IPFIX concentrators,” draft-kobayashi-ipfix-concentrator-model-01.txt (work in
progress).
9. R. Enns, “NETCONF Configuration Protocol,” draft-ietf-netconf-prot-12.txt (work in
progress).
10. T. Dietz and B. Claise “Definitions of Managed Objects for Packet Sampling,” draft-ietf-
psamp-mib-05.txt (work in progress).
11. A. Kobayashi, K. Ishibashi, K. Yamamoto and D. Matsubara, “Managed Objects of IPFIX
concentrator,” draft-kobayashi-ipfix-concentrator-mib-01.txt (work in progress).
12. http://silktools.sourceforge.net/
13. http://nfdump.sourceforge.net/
14. D. Abadi et al., “Aurora: A Data Stream Management System,” VLDB Journal Vol.12,
No.2, pp. 120-139, 2003.
Novel Traffic Measurement Methodology for High
Precision Applications Awareness in Multi-gigabit
Networks
1 Introduction
In recent years, traffic measurement and analysis studies have received significant
attention due to the requirements from various aspects such as SLA monitoring, P2P
traffic monitoring, and detection of security anomalies, etc. However, technical
challenges have been increased simultaneously because of the multi-gigabit link
speeds, huge traffic volume to measure, and the dynamic characteristics of the current
and newly emerging Internet applications.
High speed and volume traffic measurement requires improvements from both
hardware and software. Even packet capture in one gigabit speed using a commercial
NIC (Network Interface Card) causes significant packet losses when bursty traffic
comes in. Dedicated special hardware has to be designed to meet such high speed and
volume measurement requirements. Major capabilities that have to be considered in
the hardware are high speed accurate packet processing including fragmented packet
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 63 – 72, 2006.
© Springer-Verlag Berlin Heidelberg 2006
64 T. Choi et al.
2 Related Work
There have been many research and development efforts in the field of traffic
measurement and analysis for the past decade. As a result, many tools were
Novel Traffic Measurement Methodology for High Precision Application Awareness 65
3 Methodology
In this section, we propose our novel methodology to fulfill the requirements in terms
of hardware and software.
In comparison with the existing traffic capture cards which were described in the
section 2, we have designed ours with more strict objectives to meet the above
requirements: packet captures without loss at upto multi-gigabit speed, lowest CPU
resource utilization as possible, filtering, packet and flow sampling, deep packet
inspection, anomaly detection, and flow generation.
Based on the design, we have developed a series of capture cards: DS-3, Fast
Ethernet, OC-3 ATM/POS, OC-12 ATM/POS, and Giga Ethernet for the lower and
medium speeds. We are also working on a metering system for 2.5 and 10 Gbps
speeds. We have decided to develop a standalone system rather than a PCI-type card
66 T. Choi et al.
due to various strong requirements we have set. It can not only capture the packets
upto 10 Gbps speed without loss but also support wire-speed deep packet inspection
for applications signature detection, flow sampling, anomaly traffic detection and
special handling to relieve measurement burden, and, most importantly, flow record
generation in the hardware. Our flow record format extends that of the netflow type
records to support high precision applications analysis.
As mentioned above, packet capture cards that support upto 10 Gbps link speed
have been developed. However, their capability is limited for packet header
processing level. Any host system which installs the capture card can’t support its
wire-speed performance for the analysis work. Especially, when it comes to flow
records generation, it represents very limited performance. It is mainly due to the fact
that flow creation is usually done at the software level. In order to catch up with such a
packet capturing speed, most of time consuming processes have to be pushed into the
hardware level. Such attempt hasn’t been tried by any researches before due to its high
complexity and cost. It is especially challenging for us that we are trying to conduct
hardwired flow creation with some extensions to achieve precise applications traffic
accounting.
tGz
h G{GtGt mGnGt
k
wGjGjGp z
l
t
wGjGt
capabilities. It supports static, probabilistic, and flow sampling. And for precise
application traffic recognition and anomalous traffic detection, we added content
inspection capability in it. It supports up to 16 bytes payload searching.
The novel mechanisms in our traffic measurement and analysis software consist of the
following four items: general purpose Internet applications traffic classification,
accurate application accounting, adaptability and expendability for newly emerging
applications, and auto-detection mechanism of the new applications.
Most application monitoring systems currently available focus on a specific target
such as P2P and/or streaming applications identification and security anomaly
detection. Of course, there is an appropriate rationale for each effort. Our initial goal
is the precise Internet application accounting for charging. Thus, we have put our
efforts to come up with a general purpose Internet application identification
methodology. Since we are using several novel methods which are described in details
below to identify most of the popular Internet applications, the accuracy of application
traffic accounting is much higher in comparison with most other currently available
monitoring solutions which depend on port to application mapping. Since Internet
applications lifecycle is very dynamic, it is very important to design the monitoring
system to adapt such a characteristic. We added a run-time configuration language for
such a purpose. When a new application appears in the Internet, we need to identify it
promptly for its monitoring. Manual detection is very time-consuming and labor
intensive work and thus requires automation. We are currently working on such
automation and a preliminary result is explained below. Our software methodology
consists mainly of a precise application identification method, extensible applications
recognition language, and flow definition extension.
registered ports can be simply recognized by this method. There exists, however, a
rather higher probability of misrecognition due to the current Internet applications
characteristics as explained before.
- Type-PI (Payload Inspection-based Recognition Type): Recognition is performed
on the basis of both port numbers and signatures, a.k.a. patterns, in the application
PDU (Payload Data Unit). This method produces an effect when two or more
equivalently influential applications share a registered port number. Any well known
services or popular applications can also be identified by this method if higher level of
correctness assurance is required.
- Type-DP (Dynamic Port-based Recognition Type): Recognition is performed on
the basis of port numbers obtained by inspecting other flows’ payloads. In the sense of
payload inspection, this method is similar to type-PI; however, the difference is that, in
this type, the sought pattern provides a referential hint to identify another flow (a type-
DP flow or an induced flow) that may take place soon after. One of the common type-
DP flow examples is a passive mode FTP.
- Type-RR (Reverse Reference-based Recognition Type): Recognition is performed
on the basis of a referential note obtained by recognizing a type-PI flow on the other
links. We define a reverse flow: When there exists a flow, X, of which <src_addr,
dst_addr, src_port, dst_port, protocol> is <a, b, x, y, p>, if another flow, Y, is specified
by (b, a, y, x, p), then Y is a reverse flow of X. The purpose of the type-RR method is,
thus, to recognize reverse flows of type-PI flows. In most cases, type-RR flows are
control flows of legitimate TCP connections whose constituting packets contain only
IP and TCP headers or flows whose constituting packets do not contain any distinctive
patterns in the payload
4 System Architecture
payload portion of a packet which is matched by the signature. Our extended flow
records, then, are sent to the Collector. It aggregates all the flow records collected
from the multiple metering systems and export them to the Analysis server in real-
time or store them in the Storage server for near real-time processing. The Analysis
server classifies applications based on our classification algorithm and stores the
result statistics in the DB. GUI then accesses it for various presentations.
h G n|pV~
z z
hyjs
j
m
tG OyTG
z mGkP
j
zG
ki
OmGGwGkP
zGz
upto 1 Gbps and the other for higher speed upto 10 Gbps. The former system has
been installed and operated on a number of sites including a large scale campus
network, a medium scale enterprise network (ETRINet), KIX - one of the biggest
public Internet exchange (IX), and some other ISP’s international 1 Gbps links in
Korea. The latter system is under development and the prototype will be available in
the third quarter of 2006 for the testing.
For the former, we have implemented intelligent packet capture (IPCAP) cards
suited for our specific architecture and needs, they include DS-3, FastEthernet, OC-
3/12 ATM/POS, and GigaEthernet Cards. They are PCI type-II based cards.
Currently, we verified that cards upto 1Gbps can support wire speed packet capture
without a single packet loss even for 64 byte IP packets arriving back-to-back.
In ETRINet, we monitored the most aggregated link which connects ETRINet to
two major ISPs in Korea via two T3 (45 Mbps) links. The number of Internet users in
ETRINet reaches around 2,500. Outgoing and incoming traffic is simultaneously fed
into a Metering system and to a Server from a mirror-enabled Ethernet switch. The
average profile of the consolidated traffic is 46.52 Mbps, 5.325 Kpps, and 174.1 fps,
and the peak profile is 148.23 Mbps, 18.242 Kpps, and 1.359 Kfps. Two logical links
(outgoing and incoming) are monitored by a single metering system. The Collector
and the Server platform is composed of dual Pentium-IV CPUs, 2 GB of memory, and
a 66MHz 64-bit PCI bus respectively.
By using our IPCAP GigaEthernet cards in ISP’s international link, we have
observed 0% packet loss on both incoming and outgoing links during the period of 6
weeks in March and April of 2006. Out-going link has around 200Mbps and incoming
link has around 180Mpbs traffic utilization. Successful application recognition ratio
is around 85% and we are trying to improve the unknown ratio. We have also tested
POS OC-3 cards in another major Korean ISP’s international link. This link was fully
utilized and extensive system improvement and optimization work has been
conducted. Thus, OC-3 card based monitoring system is recently commercialized.
We are attempting various tests with OC-48 and OC-192 systems until the fourth
quarter of 2006 and the performance analysis results can be incorporated in the final
version of this paper.
Due to highly dynamic nature of the development and the use of the current Internet
applications, accurate application traffic usage accounting in the Internet requires a
cleverly combined mechanism of per-packet payload inspection, flow-based analysis,
correlation of associated sub-transaction flows, and wire-speed packet capturing
performance. In this paper, we proposed the novel approach to meet such challenges.
In addition to designing architecture and proposing technical solutions, we have
embodied them in a series of Wise*TrafView systems. We are satisfied with the initial
experiences with our systems.
We have tested our system with upto 1Gbps speed and are currently working for
enhancing it to support much higher speeds such as OC-48 & OC-192. It is very
72 T. Choi et al.
challenging that this new system shifts major functionality into hardware level. As
far as we understand, such attempt hasn’t been made by any researches yet. We are
expecting to release our next version by the fourth quarter of 2006. Although we
have focused on the usage-based accounting in this paper, it can be utilized in many
other areas such as traffic profiling and security anomaly detection. These additional
capabilities will be explored as our future work as well.
References
1. Colleen Shannon, David Moore, and k claffy: Characteristics of Fragmented IP Traffic on
Internet Links. Proc. of ACM SIGCOMM Internet Measurement Workshop, San
Francisco, USA, Nov. 2001
2. CIADA’s OCxMon & NetTraMet. http://www.caida.org/tools/
3. TCPDUMP. http://sourceforge.net/projects/tcpdump/
4. Ethereal. http://www.ethereal.com/
5. Sprint ATL, "IP Monitoring Project,“ http://www.sprintlabs.com/Department/IP-
Interworking/Monitor/
6. Http://www.cisco.com/univercd/cc/td/doc/cisintwk/intsolns/netflsol/nfwhite.htm.
7. K. Keys, D. Moore, Y. Koga, E. Lagache, M. Tesch, and K. Claffy: The Architecture of
CoralReef: An Internet Traffic Monitoring Software Suite. Proc. of Passive and Active
Measurement Workshop 2001, Amsterdam, Netherlands, April 2001
8. D. Plonka, Flowscan: A network traffic flow reporting and visualization tool. In
Proceedings of USENIX LISA, 2000
9. DAG cards, Endace Measurement Systems, http://www.endace.com
10. Jan Corppens, et. al., “SCAMPI – A Scaleable Monitoring Platform for the Internet”,
Technical Report, EU IST project, 2003
11. Cisco NBAR. http://www.cisco.com/warp/public/732/Tech/qos/nbar/
12. Jacobus van der Merwe, Ramon Caceres, Yang-hua Chu, and Cormac Sreenan “mmdump-
A Tool for Monitoring Internet Multimedia Traffic,” ACM Computer Communication
Review, 30(4), October 2000.
13. Hun-Jeong Kang, Myung-Sup Kim and James Won-Ki Hong, "A Method on Multimedia
Service Traffic Monitoring and Analysis", Lecture Notes in Computer Science 2867,
Edited by Marcus Brunner, Alexander Keller, 14th IFIP/IEEE International Workshop on
Distributed Systems: Operations and Management (DSOM 2003), Heidelberg, Germany,
October, 2003, pp. 93-105.
14. Subhabrata Sen and Jia Wang, "Analyzing peer-to-peer traffic across large networks", in
Proceedings of the second ACM SIGCOMM Workshop on Internet Measurement
Workshop, Nov. 2002.
15. Alexandre Gerber, Joseph Houle, Han Nguyen, Matthew Roughan, and Subhabrata Sen,
"P2P The Gorilla in the Cable", National Cable & Telecommunications Association
(NCTA) 2003 National Show, Chicago, IL, June 8-11, 2003.
16. Nathaniel Leibowitz, Matei Ripeanu, and Adam Wierzbicki, "Deconstructing the KaZaA
Network", 3rd IEEE Workshop on Internet Applications (WIAPP'03), June
17. http://www.cisco.com/en/US/products/ps6151/index.html
18. http:// www.netintact.com/
19. http:// www.picmg.org/newinitiative.stm
Rate-Based and Gap-Based Available Bandwidth
Estimation Techniques in Cross-Traffic Context
1 Introduction
1
A(t − τ , t ) = C (1 − u ( x ))dx ,
t
(1)
τ ³ t −τ
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 73 – 81, 2006.
© Springer-Verlag Berlin Heidelberg 2006
74 W. Tan, M. Zhanikeev, and Y. Tanaka
where C is the capacity of the link and u(x) is the instantaneous utilization of the link at
time x. Similarly, the end-to-end available bandwidth is determined by the smallest
link available bandwidth in the path over certain interval.
End-to-end available bandwidth is usually estimated with active probing techniques
which send probe traffic from the source host to the destination host of the path. Active
probing is free of privileged access requirement and feasible for the end users. Various
active probing techniques have been proposed today. Generally, they can be classified
into two models according to the underlying approaches:
The rate-based model (RM) uses the sending rate of the probe traffic at the sender
(probing rate) to infer end-to-end available bandwidth. The RM techniques usually
send probe packets from an initial low rate and increase the rate gradually. They search
for the turning point at which the arrival rate of probe packets at the receiver strata to be
lower than the probing rate. Such turning point is believed to mirror the end-to-end
available bandwidth over the probing period. RM techniques include pathChirp [1],
Pathload [2] and PTR [3].
The gap-based model (GM) compares the time gap of successive probe packets
between the sender and the receiver to calculate the available bandwidth. Given two
successive probe packets fall in the same queuing period at the single bottleneck of the
path, and they are sent with a time gap ǻin and reach the receiver with a time gap ǻout,
then ǻout should consist of two segments: ǻin and the transmission time of the
cross-traffic between them. The GM techniques use the difference between ǻout and ǻin
and end-to-end capacity to calculate the rate of cross-traffic and then end-to-end
available bandwidth. The calculation of GM tools is usually based on a set of
congested probe packets. IGI [3] and Spruce [4] are examples of tools using GM.
In this section two representative tools of each class are selected to be compared. In the
RM class pathChirp [1] is famous for its chirp structure – a series of exponentially
spaced packets within a single probe train. The chirp structure covers a wide range of
probing rate and thus works with high efficiency. The probing process of pathChirp is
fast and with significantly light overhead traffic. IGI [3] is a tool in the GM class. IGI
sends an even spaced train with carefully selected initial gap and increases the gap
linearly for subsequent trains. IGI estimates available bandwidth when all the gaps of
train at the sender equal to those at the receiver. PathChirp and IGI are selected to
represent their class in our study.
Some evaluation of the existing end-to-end available bandwidth estimation tools has
been done on high speed links (with capacity of 1000 Mbps) environment [5].
Rate-Based and Gap-Based Available Bandwidth Estimation Techniques 75
However, high speed end-to-end path does not yet prevail in today’s Internet; most
paths are with a capacity around 10 Mbps, usually constrained by the edge of the
network. For practical purpose, we evaluate and compare the tools on paths with
capacity of 10 Mbps. The test is performed based on simulations with OPNET
Modeler. Simulation environment is crucial for this research as we need complete
control over cross-traffic load on network paths and the ability to monitor link
utilization with high level precision and granularity, which is not possible in real
network environment. The network paths under observation are from 5 to 7 hops long
with a single bottleneck in the middle of the path whose capacity is 10 Mbps.
Admittedly, this topology is primitive and some special artifacts such as multiple
bottlenecks are beyond the scope of our consideration, but the simple topology is
favorable for discovering the fundamental characteristics of the tools.
Most available bandwidth estimation tools work well in stable cross-traffic
environment. When they face highly bursty cross-traffic, however, the performance is
very different for each tool. Therefore the performance in bursty cross-traffic
environment is critical for evaluation. Except the idle path and light traffic scenarios,
most of our test scenarios simulate highly bursty traffic environment. The cross-traffic
is generated in the form of various popular Internet applications such as HTTP web
browsing, FTP file transferring, Email, video conferencing and so on.
Both pathChirp and IGI have tunable parameters. Generally the default values of
those parameters are used in our test. The initial gap of IGI is exceptional, because
some test scenarios with high utilization are performed and the default initial gap is too
large. For pathChirp, the probe packet size is 1000 bytes, the exponential spread factor
is 1.3 and the length of the chirp is 15; for IGI, the probe packet size is 500 bytes, the
initial gap is 0.5 ms and the length of a train is 32. This setting remains unchanged for
all the comparison scenarios. Two tools run subsequently on the same path with the
same cross-traffic condition repeated by the simulator.
The first test scenario is performed on an idle path. The results are shown in Fig. 1.
Both tools give an estimate around 10 Mbps, which is the end-to-end capacity of the
path. While the estimate of IGI is very close to 10 Mbps, it is interesting to see that the
estimate of pathChirp goes up to 11 Mbps. When pathChirp is evaluated on high speed
links [5] the same phenomenon is also seen. Please note that the actual available
bandwidth (avail-bw for short in the figures) is not 10 Mbps in both pathChirp and IGI
cases. It is because the probe traffic occupies a little bandwidth of the path.
In the second scenario, there is light cross-traffic with average utilization at about
15%. The results of both tools, which are also shown in Fig. 1, are exactly the same as
those on an idle path. It could be understood, because 15% of 10 Mbps is not more than
the probe traffic rate of both tools. It means the cross-traffic is extremely light.
Actually in a path with capacity of 10 Mbps both tools can not sense cross-traffic when
the average utilization is lower than 15%.
Fig. 2 shows the results when the average utilization is 30%. Due to the bursty
cross-traffic, the instantaneous utilization is heavily fluctuating. The accuracy of both
76 W. Tan, M. Zhanikeev, and Y. Tanaka
tools drops in this scenario. However, the estimate results of both tools show different
characteristics. The estimate of pathChirp exhibits highly dynamic feature which
reflects the fluctuation of the cross-traffic. A number of estimate results correctly
follow the change of cross-traffic although some go opposite direction. But pathChirp
is most of time over-reacting: even it follows the change of cross-traffic but the
estimate value is either too high when the available bandwidth increases, or too low
when it decreases. Some estimate results are higher than 10 Mbps which is obviously
over the border. On the other hand, IGI barely corresponds to the change tendency of
the cross-traffic. The range of the estimate results is limited. In this scenario, they are
almost always higher than actual available bandwidth and fall into the area between
actual available bandwidth and end-to-end capacity.
There is more cross-traffic in the next scenario and the average utilization is 50%.
Fig. 3 shows the results in this scenario. Again pathChirp shows its strong ability to
follow the change of cross-traffic but is over-reacting. The performance of IGI is also
unchanged.
The final scenario is with heavy cross-traffic and the average utilization jumps up to
70%. Fig.4 shows the results in this scenario. Unsurprisingly, the particular
characteristics of both tools exist as the same. But, pathChirp gives less estimates
lower than the actual available bandwidth compared with previous scenarios. And,
there are occasionally extremely high estimate results up to 30 Mbps from pathChirp.
In the above test scenarios, pathChirp and IGI exhibit different characteristics:
pathChirp is good at following changes of cross-traffic but often over-reacts, while IGI
does not catch changes of cross-traffic accordingly but offers a stable estimate. The
different characteristics probably result from the underlying approaches of two models.
The RM tools perform estimation based on a single turning point. For pathChirp, the
turning point is a single packet pair inside the chirp. Therefore, it is fast to discover the
change of cross-traffic but easy to deviate from the actual value. The GM tools estimate
available bandwidth based on a set of congested probe packets. The estimate is
smoother but can not follow changes in cross-traffic promptly.
Fig. 1. Comparison between pathChirp and IGI without cross-traffic and with light cross-traffic
Rate-Based and Gap-Based Available Bandwidth Estimation Techniques 77
Fig. 2. Comparison between pathChirp and IGI with average utilization at 30%
Fig. 3. Comparison between pathChirp and IGI with average utilization at 50%
Fig. 4. Comparison between pathChirp and IGI with average utilization at 70%
78 W. Tan, M. Zhanikeev, and Y. Tanaka
We evaluate the hybrid method by comparing it with pathChirp and IGI. The
comparison is performed in the same environment and with the same methodology as
previous tests.
First, the hybrid method is compared with pathChirp and IGI when the average
utilization of the path is 30%. The results are shown in Fig. 5. The coefficient Į is 0.5
in this case. The hybrid method yields much closer results to the actual available
bandwidth than the other two tools. It can not only follow changes of cross-traffic but
Rate-Based and Gap-Based Available Bandwidth Estimation Techniques 79
also effectively restrain the estimates from overreaction. In addition, the hybrid
method reduces probe traffic in the second phase compared with the original IGI. The
average number of trains of process in the hybrid method is 1.7 while the number is
6.2 for original IGI.
Fig. 6 shows the comparison when the average utilization is 50%. The coefficient Į
is also 0.5. Again the hybrid method performs the best. The average number of
trains is 3.5 and that of original IGI is 9.1.
The final comparison is performed when the average utilization is 70%. The results
are shown in Fig. 7. In this scenario, the coefficient Į is 0.7. The hybrid method is still
the best among the three. The average number of trains is 3.7 and that of original
IGI is 7.6.
We set the coefficient Į with different values to see how it affects the final estimate.
The error rate is used for verification. The error rate is the difference between the
estimate and the actual available bandwidth value in proportion to the total capacity. A
single error rate value is calculated from a number of estimate
Fig. 5. Comparison among hybrid method, pathChirp, and IGI with average utilization at 30%
Fig. 6. Comparison among hybrid method, pathChirp, and IGI with average utilization at 50%
80 W. Tan, M. Zhanikeev, and Y. Tanaka
Fig. 7. Comparison among hybrid method, pathChirp, and IGI with average utilization at 70%
samples of consecutive probes. The results are shown in Fig. 8. When the load is not
heavy (e.g., the average utilization is under 50%), the value of 0.5 yields the best result.
When the load is very heavy, the value of Į should also increase because the IGI
estimates are generally high and they should be restricted to a light-weight. Generally
speaking, a network path with normal usage should not see very heavy utilization, so
0.5 is appropriate for the coefficient value. When heavy utilization happens, however,
we can refer to the estimates of IGI as it is stable. For example, we can assume the
utilization is heavy when the average estimate of IGI is no more than 6 Mbps in our test.
This assumption is based on our empirical results.
4 Conclusion
In this paper the differences in performance between rate-based model and gap-based
model of end-to-end available bandwidth estimation techniques are studied. PathChirp
and IGI are selected to represent each model. The evaluation is performed on low
speed paths which are typical in today’s Internet. A hybrid method adopting both
Rate-Based and Gap-Based Available Bandwidth Estimation Techniques 81
References
1. Ribeiro, V. J., Riedi, R. H., Baraniuk, R. G., Navratil, J., and Cottrell, L.: pathChirp: Efficient
vailable Bandwidth Estimation for Network Paths. Proc. Passive and Active Measurement
Workshop, Tech. Report no. SLAC-PUB-9732 (2003)
2. Jain, M., and Dovrolis, C.: End-to-End Available Bandwidth: Measurement Methodology,
Dynamics, and Relation with TCP Throughput. IEEE/ACM Trans. Networking, Vol.11, No.4
(2003) 537-549
3. Hu, N. and Steenkiste, P.: Evaluation and Characterization of Available Bandwidth Probing
Techniques. IEEE J. Select. Areas Commun., Vol.21, No.6 (2003) 879-894
4. Strauss, J., Katabi, D., and Kaashoek, F.: A Measurement Study of Available Bandwidth
Estimation Tools. Proc. 3rd ACM SIGCOMM Conf. Internet Measurement (2003) 39-44
5. Shriram, A., Murray, M., Hyun, Y., Brownlee, N., Broido, A., Fomenkov, M., and Claffy, K.:
Comparison of Public End-to-End Bandwidth Estimation Tools on High-Speed Links. Proc.
Passive and Active Network Measurement, 6th International Workshop, Boston, MA, USA
(2005)
Signature-Aware Traffic Monitoring with IPFIX*
Abstract. Traffic monitoring is essential for accounting user traffic and detecting
anomaly traffic such as Internet worms or P2P file sharing applications. Since
typical Internet traffic monitoring tools use only TCP/UDP/IP header information,
they cannot effectively classify diverse application traffic, because TCP or UDP
port numbers could be used by different applications. Moreover, under the recent
deployment of firewalls that permits only a few allowed port numbers, P2P or
other non-well-known applications could use the well-known port numbers.
Hence, a port-based traffic measurement scheme may not provide the correct
traffic monitoring results. On the other hand, traffic monitoring has to report not
only the general statistics of traffic usage but also anomaly traffic such as
exploiting traffic, Internet worms, and P2P traffic. Particularly, the anomaly
traffic can be more precisely identified when packet payloads are inspected to find
signatures. Regardless of correct packet-level measurement, flow-level
measurement is generally preferred because of easy deployment and low-cost
operation. In this paper, therefore, we propose a signature-aware flow-level traffic
monitoring method based on the IETF IPFIX standard for the next-generation
routers, where the flow format of monitoring traffic can be dynamically defined
so that signature information could be included. Our experimental results show
that the signature-aware traffic monitoring scheme based on IPFIX performs
better than the traditional port-based traffic monitoring method. That is, hidden
anomaly traffic with the same port number has been revealed.
Keywords: signature, IPFIX, traffic measurement, flow, and security.
1 Introduction
Traffic monitoring is essential for accounting normal user traffic and detecting
anomaly traffic such as Internet worms or P2P file-sharing applications. In general,
simple packet- or byte-counting methods with SNMP have been widely used for easy
and simple network administration. However, as applications become diverse and
anomaly traffic appears quite often, more detailed classification of application traffic
is necessary.
*
This research was supported by the MIC (Ministry of Information and Communication),
Korea, under the ITRC (Information Technology Research Center) support program
supervised by the IITA (Institute of Information Technology Assessment). (IITA-2005-
(C1090-0502-0020)).
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 82 – 91, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Signature-Aware Traffic Monitoring with IPFIX 83
2 Related Work
Typically, flow-level traffic measurement was done with Cisco NetFlow. FlowScan
[3], that generates and visualizes traffic with NetFlow, uses port numbers for
classifying applications. However, port-based traffic classification methods may be
incorrect, because port numbers could be used by other applications. Although
packet-level traffic measurement [4] could generate more precise results, it is
expensive and difficult to deploy in a large-scale network.
In general, snort [5], which is a widely-used open IDS, can detect anomaly traffic
such as Internet worms, viruses, or exploiting incidents including signatures. Thus,
alert messages and logs are sent and recorded. However, the purpose of the IPS is to
detect anomaly traffic.
Recently, a few content-aware traffic monitoring methods [6][7] have been
proposed. In [6], signatures were used to classify traffic for accounting, and it was
shown that traffic of well-known ports includes that of non registered applications.
However, it does not support IPFIX. In [7], various traffic classification methods
including packet inspection have been compared, and it was explained that unknown
traffic could be correctly identified through searching signatures of the first packet,
the first a few Kbytes, a few packets, or all the packets of the flow. However, these
two studies use their own proprietary architectures for traffic measurement. In this
paper, we propose a signature-aware traffic monitoring scheme that employs the
IPFIX standard which could be used by next-generation routers.
3.1 Architecture
1
Optionally, UDP may be used.
86 Y. Lee, S. Shin, and T.-g. Kwon
aggregate many flows into a simplified form of flows. Since a lot of flow data are
continuously exported to the flow collector, a post-end database system is integrated
with the flow collector for further analysis.
Every IPFIX message consists of an IPFIX message header, a template set, and a data
set (an option template set and option data set) as shown in Fig. 2. A template set
defines how the data set is organized. A newly created template is sent through an
IPFIX message consisting of interleaved template set and data set (option template set
and option data set). After the template set has been delivered to the IPFIX collector,
following IPFIX messages can be made up with only data sets. When UDP is used as
the transport protocol, template records and option template records must be
periodically sent.
We defined a new flow template set including the signature ID field2 as shown in
Fig. 3-(a). The FlowSet ID of 0 means that this flow is the template. Basically, the
flow defined by the template in Fig. 3-(a) delivers bytes, packets, flow start/end time,
and signature ID for a flow of (src IP, dst IP, src port, dst port, protocol). Here, we use
the signature ID values same with snort. Therefore, if the signature inspector finds a
signature, it will record the signature ID at this field.
In Fig. 3-(b), the real example of the IPFIX data set which complies with the IPFIX
template in Fig. 3-(a) is shown. The Template ID (=256) in Fig. 3-(a) and the FlowSet
ID (=256) should be same if the flow data record is to be parsed according to the
given template set. The web flow between 168.188.140.87 and 211.115.109.41 has
3,482 bytes, 5 packets, and the signature ID of 1855 which is related with the DDoS
attack. Generally, in a single flow packet, more than one flow data set will be
contained.
2
The type of the “signature ID” is defined to 200 and the length of the “signature ID” is 2
bytes.
Signature-Aware Traffic Monitoring with IPFIX 87
Fig. 3. IPFIX template and flow data message format including signature ID
4 Experiments
Inbound Outbound
Total bytes 3.2TB 2.4TB
Total packets 6,812,926,748 7,272,913,398
Total flows 65,130,555 80,017,160
Inbound Outbound
Signature ID Number of flows Signature ID Number of flows
528 278,574 528 271,781
483 34,064 30004 255,711
525 19,338 1855 27,890
485 12,669 525 17,575
1419 9,454 1419 12,316
312 8,880 480 11,549
1417 8,871 2586 11,448
1201 4,532 312 8,486
1200 2,460 1417 8,373
486 2,162 2181 6,480
The detailed per-port statistics are shown in Table 3. In inbound traffic, various
ICMP-based attack patterns have been found at port 0. Similarly, signatures are
observed at well-known ports of 20, 22, 80, and 8080 as well as not-well-known ports
of 2420, 2725, 3389, 4075, and 5063. In outbound traffic, one interesting port is
19101 which is used for web disk service of exchanging files.
Signature-Aware Traffic Monitoring with IPFIX 89
Inbound Outbound
% of
Total % packets
Destination packets Destination Total number
number of with
Port with Port of packets
packets signatures
signatures
0 198,190,150 2.9 80 1,446,312,138 20.0
80 180,441,016 2.6 19101 338,463,084 4.7
20 141,035,416 2.1 8080 329,413,874 4.5
8080 48,638,816 2.1 7132 290,554,498 4.0
2420 48,638,816 0.7 5090 273,500,608 3.8
2725 18,907,224 0.3 7778 182,935,034 2.5
5063 17,004,268 0.3 0 171,201,682 2.4
3389 16,867,212 0.2 23 154,077,628 2.1
4075 15,295,958 0.2 5004 140,164,098 1.9
22 14,619,240 0.2 6699 113,544,400 1.6
At the specific port number, the found signature information is widely distributed.
For example as shown in Fig. 5, “BitTorrent” signature 2180 has been found in
outbound link. In addition, at port 80, other signatures such as “bad traffic with
loopback address (528)”, “web-misc whisker tab splice attack (1087)”, “spyware-put
trackware (5837)”, and “DDoS attack (1855)”. From the experiments, it was shown
that our signature-aware traffic monitoring method can illustrate the hidden P2P or
anomaly traffic patterns.
Figure 6 is a snapshot of our tool [9] which can visualize signature-aware IPFIX
flow data exported from routers. The traffic with signatures of 527 and 2586 has been
shown. The signature ID of 527 is related with a DoS traffic attack with the same
source and destination addresses. The signature ID of 2586 is the eDonkey traffic
which has “E3” signature in the payload as follows.
5 Conclusion
References
[1] Cisco NetFlow,
[2] J. Quittek, T. Zseby, B. Claise, and S. Zander, “Requirements for IP Flow Information
Export (IPFIX),” IETF RFC3917, Oct. 2004.
[3] D. Plonka, “FlowScan: A Network Traffic Flow Reporting and Visualization Tool,”
USENIX LISA, 2000.
[4] C. Fraleigh, S. Moon, B. Lyles, C. Cotton, M. Khan, D. Moll, R. Rockell, T. Seely, and C.
Diot, “Packet-Level Traffic Measurements from the Sprint IP Backbone,” IEEE Network,
vol. 17 no. 6, pp. 6-16, Nov. 2003.
[5] M. Roesch, “Snort - Lightweight Intrusion Detection for Networks,” USENIX LISA, 1999.
[6] T. Choi, C. Kim, S. Yoon, J. Park, B. Lee, H. Kim, H. Chung, and T. Jeong, “Content-
aware Internet Application Traffic Measurement and Analysis,” IEEE/IFIP Network
Operations & Management Symposium, 2004.
[7] A. Moore and K. Papagiannaki, “Toward the Accurate Identification of Network
Applications,” Passive and Active Measurement Workshop, April 2006.
[8] nProbe, http://www.ntop.org/
[9] WinIPFIX, http://networks.cnu.ac.kr/~winipfix/
Temporal Patterns and Properties in
Multiple-Flow Interactions
1 Introduction
The model of self-similarity is well established in traffic analysis literature. First
parameterized models of aggregated traffic based of self-similarity can be found
in [1] and [9], and were even considered for new queuing and routing disciplines
[6]. The model was verified using LAN and high-speed WAN traffic in [7].
As the attention is mostly focused on flows at the tail of the distribution,
many research works attempt to detect only large-volume traffic. Crovella et al.
[8] developed the AEST tool that is based on self-similarity and finds a point
in the distribution where heavy tail begins. The AEST tool was used by many
researchers to extract large-volume flows [3] [4] [5].
First attempts to find correlation among flow metrics was made in [10], where
correlation of size and rate of flows was sought. Zhang et al. in [10] found strong
correlation between size and rate, as well as additional temporal characteristics,
such as the fact that correlation grew stronger in larger intervals.
Practical results of large-volume traffic detection using heavy-tail model ex-
posed the issue of temporal volatility of elephants [2]. The term volatility refers
to the fact that elephant flows may not always appear at the top list which
makes it difficult to provide steady identification of elephant flows. This artifact
was studied in detail in [5] and a method called “latent heat” was introduced in
order to allow certain freedom to elephant flows as long as temporal gaps were
not too large.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 92–101, 2006.
c Springer-Verlag Berlin Heidelberg 2006
Temporal Patterns and Properties in Multiple-Flow Interactions 93
320
318 8
316 6
Flo
Flo
ws
314 80 80
ws
312 60 2 60
40 ds ) 40 ds )
310 20 s e c on 0 20 e c on
(
Tim
e e (s
Tim
a) Top 80% flows by size in WAN b) Top 80% flows by size in LAN
In this paper a flow is attributed with a parameter that stands for the intensity
of network interactions of an application or a user at the source. Hereinafter
this parameter will be referred to as interactivity parameter. Applications that
heavily rely on feedback from flow destinations would have low interactivity,
while file transfers and otherwise generally non-stop transmissions creating one-
time large bulks of data are considered to be more interactive. Naturally, human
response would have lowest possible interactivity. The paper proves that elephant
flow volatility can be explained with the introduction of interactivity parameter
at the source. Additionally, the noval concept of interactivity allows to depart
from traditional approach that considers only flows found in the distribution
tail by giving a means to separate flows into groups based on principle flow
metrics.
1
interactivity
heavy-tail
Fraction of bandwidth
heavy-tail
3
a1 a2 a3 a4
2
interactivity
1
v = αtR. (1)
N
As long as total load within the interval V = i=1 vi < RT , i.e. the network
is not congested, flows are expected to compete for their share of traffic based
on the value of α parameter. For automatic applications that do not require any
feedback, α takes values close to 1, and low values stand for feedback-intensive
applications or users. Feedback is a combined parameter that accounts not only
for TCP ACK packets, but also for the time that applications may require to
process data, get user response, etc. Parametric study of α is beyond the scope
of this paper, but we perform the analysis of its physical properties and relation
to self-similarity properties.
Physical properties of interactivity in relation to the well established heavy-
tail behaviour are displayed in Fig.2 in form of bandwidth to number of flows
distribution. Upper curve is the normal heavy-tail distribution, where most of
traffic is confined within a few first flows ordered by size. Lower curve represents
distribution as it would be without heavy-tail, i.e. only with α defining results
of multiple-flow interactions. In contrast to conventional approach, we consider
range 1 to consist of not elephants, but rather short flows that do not require
any interactions with destination, that is have very high α. Those flows that are
traditionally considered elephants, i.e. P2P software, file transfer, etc., are placed
in range 2 with moderately high α. As α can vary depending on type of appli-
Temporal Patterns and Properties in Multiple-Flow Interactions 95
cations and even CPU of a particular host, range 2 is considerably wider than
range 1. Ranges 3 and 4 are populated by different large classes of applications,
such as web-browsing, mail, and others. Detailed study of classes is beyond the
scope of this paper.
To verify the proposed model based on interactivity parameter α, real traffic col-
lected in LAN and WAN environments was analyzed. LAN traffic was collected
in the backbone of Waseda University and WAN packet traces were obtained
at a major WAN node in Japan. Naturally, LAN and WAN traces are delivered
from two different levels of aggregation. Packet traces were processed into uni-
directional IP-port pair flows with 30s timeout. Unidirectional flows are more
prone to clean classification due to the lack of ACK packets that affect packet
size distribution. 1 hour of LAN and 5 minutes of WAN traffic was used resulting
in over 6000 flows in LAN and over 300000 flows in the case of WAN. At any
given point of time 100 flows and 5000 flows in average were found in LAN and
WAN data respectively.
From 5% to 10% of all flows in both traces are confirmed to be elephant flows
in regard to data size and the traffic they carry is normally between 60% to 80%
of all size. This complies to both the theoretical establishments of self-similarity
and data analysis results of other practical research in the field.
106 106
105 105
104 104
Flows
Flows
103 103
102 102
1.0 1.0
101 0.8 101 0.8
)
0.6 0.6
(X
100 100
)
0.4 0.4
(X
n
0.8
pa
0.2 0.8 0.6 0.2
ze
Lifes 0.6 0.4
es
Size 0.4 0.2
Si
pan 0.0 0.0
(Y) 0.2 0.0
Lif
(Y) 0.0
a) Size/lifespan results in WAN
105 105
4
10 104
Flows
Flows
103 103
2
10 102
1
10 101
100 100
0.2 0.2
Lif 0.4 0.4
es 0.6
pa 0.8 0.2 0.0 Siz 0.6 0.2 0.0
n( 0.6 0.4 (X) e ( 0.8
Y) 1.0 0.6 0.4 an (X)
Y) 1.0 1.0 0.8 Size 0.8 es p
1.0 Lif
b) Size/lifespan results in LAN
106 106
105 105
104
Flows
104
Flows
103
Lifespan (Y)
103 1.0
102 0.8
1.0 102 0.6
101 0.8
0.4
)
0.6
(X
101
100 0.4 0.2
n
pa
Flows
103 103
2
10 102
101 101
100 100
0.2 Lif 0.20.4
Ra 0.40.6 es 0.6 0.0
te 0.2 0.0 pa 0.4 0.2 )
(Y 0.81.0 0.6 0.4 an (X) n ( 0.8
Y) 1.0 0.8
0.6 e (X
) 0.8 s p at
1.0 Life 1.0 R
b) Lifespan/rate results in LAN
106 106
105
105
104
Flows
104
103
Flows
103 1.0
Size (Y)
102 1.0 0.8
0.8 102 0.6
101 0.4
0.6 101
)
100
(X
0.4 0.2
0.8 0.6 100
ze
0.2 0.0
0.2 0.4 0.6 0.8
Si
Rate 0.4 0.2 0.0 1.0
(Y) 0.0 Rate (X)
a) Size/rate results in WAN
105 105
4
10 104
Flows
Flows
103 103
102 102
101 101
100 100
0.2 0.2
0.4
Ra 0.40.6 0.0 Siz 0.6 0.0
te 0.4 0.2 ) e ( 0.8 0.4 0.2
(Y 0.81.0 0.6 (X Y) 1.0 0.8
0.6 ( X )
)
1.0
0.8 Size 1.0 Rate
b) Size/rate results in LAN
1.0 1.0
Match ratio
Match ratio
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0.0 0.0
80 80
By 120 40 By 120 40
siz 160 80 siz 160 80
e 120 ate e 120
160 By r 160 ra e
t
By
a) Top N flows in WAN traffic with 1s (left) and 5s intervals
1.0 1.0
Match ratio
Match ratio
0.8 0.8
0.6 0.6
0.8 0.8
0.6 0.6
By
By
0.4 1.0 0.4 1.0
0.8
s
0.8
s
ize
ize
0.2 0.6 0.2 0.6
0.4 ate 0.4 ate
0.2 By r 0.2 By r
b) Top N flows in LAN traffic with 10s (left) and 30s intervals
Fig. 6. Cumulative match ratio between top N flows ordered by size and rate
1.0 1.0
Match ratio
Match ratio
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0.0 0.0
0.2 0.2
0.4 0.2 0.4 0.2
By 0.4 By 0.4
siz 0.60.8 0.6 siz 0.60.8 0.6
e 0.8 t e e 0.8 t e
1.0 ra 1.0 ra
1.0 By 1.0 By
a) Top % of flows in WAN traffic with 1s (left) and 5s intervals
1.0
Match ratio
1.0
Match ratio
0.8
0.8 0.6
0.4
0.6 0.8
0.8 0.6
By
By 0.6 1.0 1.0
siz 0.4 0.8 0.4
0.8
s
0.6
ize
e 0.2 0.2 0.6
0.4 0.4
0.2 rate 0.2 By r
ate
By
b) Top % of flows in LAN traffic with 10s (left) and 30s intervals
Fig. 7. Cumulative match ratio between threshold-selected top flows ordered by size
and rate
Fig.7 is the result of matching performed using thresholds to create top lists
instead of fixed number of flows. Although this figure generally supports our
argument, it also offers new properties. The perfect parallel alignment along the
size axis stands for the fact of very high variability of size, as for each threshold
of rate almost the same number of matches can be found for almost all sizes. The
smoothness of the plot also comes from the fact that temporal samples collected
over 10 successive intervals vary too much to provide any visual difference. Since,
as opposed to Fig.6, in this figure we have cases of total matches (area of low
threshold of 0.2 for each metric, i.e. all flows from the ordered list). Naturally,
all flows in that area match, thus providing the highest sample at (0.2,0.2). The
LAN part of Fig.7 displays similar pattern, with the only additional feature of
a spike at 0.2 on size axis parallel to rate, which says that when almost all flows
by size are used, they can almost always be found with any length of rate lists.
To prove higher variability in size than in rate, we performed localized match-
ing analysis in Fig.8. As opposed to previous cumulative analysis, this figure can
pinpoint the area where match ratios are higher. Thresholds are used again to
create top lists. We use the window of 0.1 to create local area and the step of
0.05 to provide smooth and detailed map.
Fig.8 contains similar features for both LAN and WAN traffic, but they are
expressed differently in each environment. WAN results clearly convey the mes-
sage that small flows by size, i.e. small number of bytes transmitted within the
interval, are often due to low transmission rate of the flow (low interactivity pa-
rameter α). This feature grows stronger in longer analysis interval. In addition,
100 M. Zhanikeev and Y. Tanaka
0.4
tio 0.4
tio
Match ra
Match ra
0.2 0.2
0.0 0.0
0.8 0.0 0.8 0.0
0.6 0.2 0.6 0.2
By 0.4 0.4 Ra 0.4 0.4
ra 0.2 0.6 ize te 0.2 0.6 e
te
0.0 0.8
By
s 0.0 0.8 Siz
1.0 1.0
a) Top % of flows in WAN traffic with 1s (left) and 5s intervals
0.8 0.8
Match ratio
Match ratio
0.6 0.6
0.4 0.4
0.2 0.2
0.0 0.0
0.8 1.0 0.8 1.0
0.6 0.8 0.6 0.8
By 0.40.2 0.6 By 0.40.2 0.6
size 0.4 e size 0.4 e
0.0 0.2 rat 0.0 0.2 rat
0.0 By 0.0 By
b) Top % of flows in LAN traffic with 10s (left) and 30s intervals
Fig. 8. Localized matches between threshold-selected flows ordered by size and rate
with the increase of the interval another feature gains strength and is positioned
at the top of size list, but in the middle of rate list. Those are elephant flows, and
the fact that they are found in the mid-area of rate axis supports the argument
that large flows are not always fast. There are no matches when both size and
rate lists are very short.
The LAN part of Fig.8 has the same features and additional peak at the area
of short size and rate lists. In this case top flows by size really do match those
by rate and the area of such matches grows with longer intervals. However, LAN
plots also have features found in WAN, such as the fact that short size lists
find good matches in lower parts of the rate list. The lack of feature when both
lists are short in WAN can be potentially explained by high variability of size,
which resulted in samples too scattered to fit in the localized matching window.
However, larger windows will result in smoother/lower matching rates, which
makes it almost impossible to verify whether there are even partial matches
at the top of both lists. In any case, we can state that high levels of traffic
aggregation result in higher variance of size, thus making the task of finding
matches more difficult.
4 Conclusion
This paper proposed a model based on source interactivity that can explain
high temporal volatility in identification of elephant flows. Based on practical
results it is displayed that elephants are not always found at the top of the list
ordered by size. The proposed source interactivity model explains this with short
Temporal Patterns and Properties in Multiple-Flow Interactions 101
traffic spikes coming from highly interactive (no feedback required) applications
that can temporarily overwhelm the list of elephants flows. In this case, the
real elephants transmit less traffic, but preserve transmission rate, which in our
model is dependent on interactivity parameter. This means that identification of
elephants by rate is more stable that that by size. This, however, is left for future
research, as well as a more detailed parameterized interactivity model that would
allow more rigid classification of traffic groups based on source interactivity, and,
subsequently, on source behaviour.
Acknowledgments
This research is sponsored by KDDI R&D Laboratories Inc.
References
1. Claffy K., Braun H. and Polyzos G.C.: A parameterizable methodology for internet
traffic flow profiling. IEEE Journal on Selected Areas in Communications 13(8)
(1995) 1481–1494
2. Abrahamsson H. and Ahlgren B.: Temporal characteristics of large IP traffic flows.
Technical Report T2003.27, Swedish Institute of Computer Science (2004)
3. Papagiannaki K., Taft N., Bhattacharyya S., Thiran P., Salamatian K. and Diot
C.: A pragmatic definition of elephants in internet backbone traffic. 2nd ACM
SIGCOMM Workshop on Internet Measurement (2002) 175–176
4. Papagiannaki K., Taft N., Bhattacharyya S., Thiran P., Salamatian K. and Diot
C.: On the Feasibility of Identifying Elephants in Internet Backbone Traffic.
Tech.Report no. RR01-ATL-110918, Sprint ATL (2001)
5. Papagiannaki K., Taft N. and Diot C.: Impact of Flow Dynamics on Traffic Engi-
neering Design Principles. IEEE INFOCOM 4 (2004) 2295–2306
6. Crovella M.: Performance Evaluation with Heavy Tailed Distributions. Lecture
Notes in Computer Science 2221 (2001) 1–11
7. Willinger W., Taqqu M., Sherman R. and Wilson D.: Self-Similarity Through
High-Variability: Statistical Analysis of Ethernet LAN traffic at the Source Level.
IEEE/ACM Transactions on Networking 5(1) (1997) 71–86
8. Crovella M. and Taqqu M.: Estimating the Heavy Tail Index From Scaling Prop-
erties. Methodology and Computing in Applied Probability 1 (1999) 55–79
9. Erramilli A., Pruthi P. and Willinger W.: Self-Similarity in High-Speed Network
Traffic Measurements : Fact or Artifact? VTT Symposium 154 (1995) 299–310
10. Zhang Y., Breslau L., Paxson V. and Shenker S.: On the characteristics and origins
of internet flow rates. ACM SIGCOMM (2002) 309–322
A Profile Based Vertical Handoff Scheme for Ubiquitous
Computing Environment
1 Introduction
With the enormous practical potential of application management and the network
manipulation method, ubiquitous computing has become a hot research area in
computer science. In the ubiquitous computing, the most important assumption is that
every component in a certain environment is connected to each other. So, seamless
connectivity must be regarded as a major challenge for the ubiquitous computing.
Specifically, in the area where heterogeneous network interface is available, it is
recommended to utilize every network interface appropriately. The management
scheme to move from one network interface to another is called vertical handoff (VHO).
Nevertheless, almost every research on VHO introduced recently is based on only
network information, such as bandwidth, signal strength, and packet error rate. And
they try to provide transparent handoff to service layer without considering other
contexts like service quality and user intention. In the aspect of application services,
any ubiquitous computing scheme not considering service quality management based
on network status or heterogeneous network environment leads to the failure of
application services. This may corrupt the continuity or stability of user intended
work.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 102 – 111, 2006.
© Springer-Verlag Berlin Heidelberg 2006
A Profile Based Vertical Handoff Scheme for Ubiquitous Computing Environment 103
2 Related Work
In this section, some related research is described. GAIA is one of the most famous
ubiquitous frameworks. It provides dynamic services based on profiles generated by
environmental contexts, e.g., network information, user moving information, and
environmental brightness. In GAIA, they suggest a scheme handling QoS when
network connection status is changed, but they did not mention how to control
network status or service quality. They just assume that every device in heterogeneous
network environments can be connected each other [1]. Aura, another famous
framework, is a good example, too. The aspect they are interested in is not to utilize
network interfaces but to manage services dynamically in a single network
environment [2].
In VHO scheme, a location service server (LSS) to provide mobile terminals with
the information about nearby networks is introduced in [3], where the VHO decision
scheme is based on the network oriented information. Thus, applications passively
receive the consequence of vertical handoff process. This mechanism causes some
unnecessary vertical handoff processes. The integration of WLAN and GSM/GPRS
with multi-tunnel scheme is proposed in [4]. This scheme uses dwell timers and
thresholds to perform the VHO properly. Although it represents QoS oriented scheme,
it does not consider any application requirement. In [6] a system for integration of
802.11 and 3G networks is proposed. The system is constructed in loosely coupled
IOTA (Integration Of Two Access technologies) gateway and IOTA client software.
IOTA gateway manages the network interfaces and prevents network oscillation, but
it only concerns network factors, such as signal strength, priority, threshold, and so
on. In [7], a new handoff protocol for overlay networks, HOPOVER (Handoff
Protocol for Overlay Networks), is proposed. HOPOVER enables smooth handoffs
104 C.-P. Hong, T.-H. Kang, and S.-D. Kim
3 Proposed Architecture
We design a profile-based middleware architecture supporting seamless connectivity
for ubiquitous computing. The proposed architecture is to use a specially designed
application profile and a profile named as working profile which represents user
intention. With the working profile and the application profile, the proposed
architecture can perform the VHO and the application service management at the
right time. In this way, we can improve the accuracy to initiate VHO and get the best
performance in network utilization. In this section, the newly designed modules and
the profiles are described in detail.
The application profile is also defined, where service mode description is a key
feature. Each service mode has different network factors. It represents the multi
modal service of an application. As network environment is changed, the application
can change its service mode based on the application profile. An example of the
application profile is presented in Table 2.
Proposed Engine
working profile is generated and passed to step 7, which facilitates the applications
service mode change. The 5th step performs the VHO decision and if necessary, finds
another proper network. Step 6 is the actual step to perform the VHO and in a certain
case it calls the application agent. The application agent manages the service mode in
the step 7 in a certain situation.
VHO decision manager decides whether it should perform any handoff and requests
the network layer to change any chosen network interface if necessary, based on the
abstracted profile and network information. The abstracted profile is generated by the
application manager and needed to specify any given network environment
supporting all the applications running on any mobile terminal. In this paper, we
assume that network layer provides the information about other available network
interfaces as well as current network interface through control channel.
1. Work
Invoc ation
2.
Applic ation
running
3. Applic ations
4. Working
Abstrac ted Profile
Available Network Exists Profile
Profile
G eneration
Generation
5. VHO
Dec ision
6. Perform
VHO
7.
Applic ation
Servic e
Mode
C hange
Abstracted Profile
ALfn AUfn
Bandwidth (Kbps) 128 500
Packet error rate (%) 3 1
Latency (ms) 20 1
normalized factors from equation (1) and (2). These equations are modified from the
expression in [5].
High factor normalization :
URm.n + LRm.n
° − 1, 2
< ALf n .
°° URm.n + LRm.n (1)
f m ,n = ®0, ALf n ≤ ≤ AUf n
° 2
° UR + LR
1, m . n m . n
> AUf n
°¯ 2
URm,n and LRm,n represent the range of real values of factor fn in the network m
and fm,n represents normalized factor in network m.
Normalized factors are updated periodically. VHO decision is activated when more
than one of average normalized factor’s value per system defined time is lower than
zero. In this manner, each factor can be considered more carefully. Thus, more precise
VHO time and QoS oriented verification can be acquired [11][13].
4 Evaluation
In this section, we present simulation results to compare our middleware architecture
with others. Single application based model without service mode management such
as AAO (active application oriented) scheme, single application based model which
A Profile Based Vertical Handoff Scheme for Ubiquitous Computing Environment 109
supports service mode management, and multi application supporting scheme which
do not support the service mode management will be employed to be compared.
When discussing the result, we use the total throughput generated by the simulation to
compare several methods. Because it denotes the total amount of information needed
to perform a certain user intended work. And also, we use the measure of the
application failures number as the reason that may cause any problem on the
continuity of the user intended work.
110 C.-P. Hong, T.-H. Kang, and S.-D. Kim
As shown in Fig. 3, the throughput of the multi application based model is higher
than single application based model by 55%. And the service mode enabled model
offers the 79% higher throughput than single application based model. Finally, the
throughput tends to increase by around 130% with the proposed scheme.
In Fig. 4, the number of application failures is presented. As shown in Fig. 4, the
number of application failures is the smallest when we apply the proposed scheme.
And the conventional scheme which is based on the single application oriented
method has the highest number of application failures. The number of application
failures for the conventional method is decreased by 85%, comparing with that of the
proposed method. It represents that the proposed scheme is an efficient method to
utilize network environment.
About 85% decreased
5 Conclusion
An efficient network management in ubiquitous computing is very important to
provide seamless service. And also, in heterogeneous network environment, a
management scheme should be designed to utilize every network interface properly.
This kind of network management scheme is called as VHO. Conventional VHO
schemes are built based only on an idea representing the transparent handoff between
heterogeneous network interfaces. And it is considered to be a different research area
to the seamless service management, so any way to provide seamless services to users
is not considered. For this reason, we design a middleware architecture with VHO
decision module, application agent, and any necessary modules. In this architecture,
we define several profiles reflecting various contexts required for seamless service.
Whit this architecture we can achieve around 130% of performance enhancement in
data throughput and around 85% of application failures can be reduced compared
with the conventional schemes.
A Profile Based Vertical Handoff Scheme for Ubiquitous Computing Environment 111
Reference
1. Román, M., Hess, C. K., Cerqueira, R., Ranganathan, A., Campbell, R. H., Nahrstedt, K.:
Gaia: A Middleware Infrastructure to Enable Active Spaces. IEEE Pervasive Computing,
Vol. 1. (2002) Page(s):74-82.
2. Sousa, J. P., Garlan, D.: Aura: An Architectural Framework for User Mobility in
Ubiquitous Computing Environments. Proceedings of the 3rd Working IEEE/IFIP
Conference on Software Architecture (2002) Page(s):29-43.
3. Inoue, M., Mahmud, K., Murakami, H., Hasegawa, M.: MIRAI: A Solution to Seamless
Access in Heterogeneous Wireless Networks. ICC2003, Vol. 2. (2003) Page(s):1033-1037.
4. Ye Min-hua, Liu Yu and Zhang Hui-min: The mobile IP Handoff Between Hybrid
Networks. PIMRC2002, Vol. 1. (2002) Page(s):265-269.
5. Chen, W. T., Shu, Y. Y.: Active Application Oriented Vertical Handoff in next-Generation
Wireless Networks. WCNC2005, Vol. 3. (2005) Page(s):1383-1388.
6. Buddhikot, M., Chandranmenon, G., Han, S., Lee, Y. W., Miller, S., Salgarelli, L.:
Integration of 802.11 and Third-Generation Wireless Data Networks. INFOCOM 2003,
vol. 1. (2003) Page(s):503-512.
7. Du, F., Ni, L. M., Esfahanian, A. H.: HOPOVER: A New Handoff Protocol for Overlay
Networks. ICC 2002, vol. 5. (2002) Page(s):3234-3239.
8. Stemm, M., Katz, R. H.: Vertical Handoffs in Wireless Overlay Networks: ACM Mobile
Networking (MONET). Special Issue on Mobile Networking in the Internet 1998. Vol. 3.
(1998) Page(s):335-350.
9. IEEE Standards Association, 802.11b Standard, http://standards.ieee.org/
10. GSM World, GPRS-Standard class 10, http://www.gsmworld.com/
11. C. P. Hong., T. H. Kang., and S. D. Kim.: An Effective Vertical Handoff Scheme
Supporting Multiple Applications in Ubiquitous Computing Environment. The 2nd
International Conference on Embedded Software and Systems (2005) Page(s):407-412
12. Kwang-Won Koh., Chang-Ik Choi., Kyung-Lang Park., Shin-Young Lim., and Shin-Dug
Kim.: A Multilayered Context Engine for the smartHome. International Conference on
Computer science, Software engineering, Information Technology, E-business, and
Applications (2004)
The Soft Bound Admission Control Algorithm for
Vertical Handover in Ubiquitous Environment
Ok Sik Yang1, Jong Min Lee1, Jun Kyun Choi1, Seong Gon Choi2,
and Byung Chun Jeon3
1 Information and Communications University (ICU),
Abstract. In this paper, we present SBAC (Soft Bound Admission Control) al-
gorithm considering critical bandwidth ratio to reduce handover blocking prob-
ability over WLAN and WAAN (Wide Area Access Network). SBAC algo-
rithm utilizes dynamically optimized resource allocation scheme to decrease the
blocking probability of vertical handover connections within the limited capac-
ity of system. Based on SBAC algorithm, we derive the handover blocking
probability as new traffic load and handover traffic load increase. In order to
evaluate the performance, we compare SBAC algorithm against traditional non-
bounded and fixed bound schemes. Numerical results show that the SBAC
scheme improves handover blocking probability in ubiquitous environment.
1 Introduction
In recent, the internet is expected to support bandwidth intensive services along with
traditional modes of data traffic with the increase of wireless devices (e.g. 3G cellular,
WLAN, Bluetooth). Furthermore, due to the increase of mobile users and environ-
mental limitation, current mobile networks need mechanisms to efficiently handle the
resource management for seamless handover in ubiquitous environment. In such envi-
ronment, a users or network will be able to decide where to handover among the dif-
ferent access technologies based on the bandwidth, cost, and user preferences, appli-
cation requirements and so on. Therefore, efficient radio resource management and
connection admission control (CAC) strategies will be key components in such a
heterogeneous wireless system supporting multiple types of applications with differ-
ent QoS requirements [1].
Many admission control schemes have been proposed to enable the network to pro-
vide the desired QoS requirements by limiting the number of admitted connections to
that network to reduce or avoid connection dropping and blocking [2], [3]. In ubiquitous
environment, other aspects of admission control need to be considered due to handover.
If the wireless network is unable to assign a new channel due to the lack of resources, an
accepted connection may be dropped before it is terminated as a result of the mobile
user moving from its current place to another during handover. Since dropping an ongo-
ing connection is generally more sensitive to a mobile user than blocking a new connec-
tion request, handover connections should have a higher priority over the new
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 112 – 121, 2006.
© Springer-Verlag Berlin Heidelberg 2006
The Soft Bound Admission Control Algorithm 113
connections in order to minimize the handover blocking probability. On the other hand,
reducing the blocking of handover connection by channel reservation or other means
could increase blocking for new connections. There is therefore a trade off between
these two QoS measures [4]. The problem of maintaining the service continuity and
QoS guarantees to the multimedia applications during handover is deteriorated by the
increase of vertical handover in heterogeneous wireless networks.
In the ubiquitous environment, vertical handover considering user preferences, traf-
fic characteristic, user mobility range, and so on could occur more frequently than
horizontal handover. Therefore, vertical handover should have higher priority to sup-
port QoS requirement because it considers more various factors (e.g. cost, bandwidth,
velocity, etc.) than horizontal handover. So we proposed a dynamic admission control
for vertical handover connections in ubiquitous environment.
This paper is organized as follows. In the next section, we describe the architecture
of proposed algorithm. In section 3, we propose a soft admission control algorithm
using softness profile. Numerical results obtained using the traditional methods are
presented and compared in section 4. Finally, we conclude the paper in section 5.
Fig. 1 shows the network architecture for mobility service in ubiquitous environment.
This architecture is based on IPv6 networks to support the movement of every user.
There are two kinds of users that can handover under this architecture. One is WLAN
to WAAN users and vice versa. Therefore, there will be two handovers between
WLAN and WAAN: vertical handover and horizontal handover.
We assume that a user has multi-interface terminal [5]. As shown in Fig. 1, the
connection initiated in WAAN and mobile node is moving to right side. When the
mobile node jumps into another access area, it requires vertical or horizontal handover
connection to continue their movement or connection, where A, B and C are handover
points respectively. A loosely coupled inter-working approach can be considered for
implementation. In this case, Mobile IP [6] mechanism must be implemented in all
equipment including MNs (Mobile Nodes). This approach also should support more
than one IP address for one mobile user so that one user can access more than one
wireless system simultaneously. Finally, on the top of a network, a suitable resource
allocation mechanism is required to control the traffic and system load. SBAC algo-
rithm is exploited here in this architecture [7].
The softness profile is defined on the scales of two parameters: satisfaction index and
bandwidth ratio [8]. The satisfaction index is a mean-opinion-based (MOS) value
graded from 1 to 5, which is divided by two regions: the acceptable satisfaction region
and low satisfaction region.
Bandwidth ratio graded from 0 to 1 can be separated by 3 regions. In the region
from 1 to A, it has low degradation of satisfaction index. It means users are not sensi-
tive in this region. However, it has large degradation of satisfaction index in the re-
gion from A to B.
The point indicated as B is called the critical bandwidth ratio ( ξ ) used in proposed
algorithm. Since this value is the minimum acceptable satisfaction index, it can be
threshold of bandwidth ratio. In the region from B to 0, users do not satisfy their ser-
vices. Therefore, this region should not be assigned to any users. Generally, the criti-
cal bandwidth ratio ( ξ ) of Video On Demand (VOD) is 0.6 ~ 0.8 and back ground
traffic is 0.2~0.6 [9].
various factors (e.g. cost, bandwidth, velocity, etc.) than horizontal handover connec-
tion. In this time, if there is no available bandwidth to accept vertical handover con-
nection, mobile agent calculates the optimized critical bandwidth ratio. If it is bigger
than threshold based on softness profile, mobile agent reassigns bandwidth based on
decided critical bandwidth ratio ( ξ ). As a result, the vertical handover connections
can be accepted more than horizontal handover connections.
Connection
Request Btotal : Total bandwidth
Boccupied : Assigned bandwidth
Bthreshold : Threshold of new connection
r : Required bandwidth
ȟ : Critical bandwidth ratio
no Is this handover no
Boccupied + r Bthreshold
connection?
yes yes
Is this vertical no
Boccupied + r Btotal
handover?
yes yes
no
ȟ threshold based on
softness profile
yes
Bandwidth reassignment
based on decided ȟ
Fig. 3 shows the system model of the proposed approach. Let C denote the total
capacity and Pnb , Phb , Pvb denote the blocking probabilities of new connection, hori-
zontal handover and vertical handover connection respectively. The arrival process of
new and handover connections is assumed to be Poisson and denoted by
λn and λh + λv . The factor α = ¬(1 − ξ ) * C ¼ is bandwidth donated from each user
116 O.S. Yang et al.
within the same coverage without critical degradation of service. From the assump-
tion, we obtain the blocking probability of horizontal handover connection, vertical
handover connection, and new connection.
Fig. 4 indicates the transition diagram for the proposed scheme. The detailed nota-
tions used in this diagram are shown in Table. 1.
Table 1. Notations
Notation explanation
λn Arrival rate of new connection
λh Arrival rate of horizontal handover connection
λv Arrival rate of vertical handover connection
1 / μh Average channel holding time for handover connections
1 / μn Average channel holding time for new connections
C Maximum number of server capacity
Threshold (bound of the total bandwidth of all accepted
T new connections)
ξ Critical bandwidth ratio
α ¬« (1 − ξ ) * C ¼»
nn Number of new connections initiated in the coverage
nhv Number of handover connections in the coverage
In order to analyze the blocking probability of each connection, we use the two-
dimensional Markov chain model with the state space S and M/M/C+Į/C+Į [10]
model is utilized.
The Soft Bound Admission Control Algorithm 117
Phb =
1
⋅¦
T
ρ nn ⋅ ρ hC − nn
(5)
P (0,0) nn = 0 nn ! (C − nn )!
1 T
ρ nn ρ h ( C +α ) − nn
Pvb = ⋅¦ ⋅
P (0, 0) nn = 0 nn ! ((C + α ) − nn )!
4 Numerical Results
In this section, we present the numerical results for the comparison of perfor-
mance. We compared three bounding schemes: non-bound, fixed bound, and SBAC
algorithms.
In Fig. 5, we increase the handover connection traffic load ( ρ h ). This graph shows
the blocking probability of handover connection under the following parameters:
C=35, T=20, λn = 1/20, λh =1/60, λ v =1/60, μ n =1/300, μ h is varying from 1/100 to
1/1000, and ξ =0.9.
In this case, traffic load of handover connections ( ρ h ) are increasing from 0 to 40
and traffic load ( ρ n ) of the new connection is 15. Since handover connections are not
bounded, as increasing the handover traffic load, the differences among three schemes
become similar.
Fig. 6 shows handover blocking probability under the following parameters: C=35,
T=15, λn = 1/30, λh =1/60, λ v =1/60, μ h =1/450, μ n is varying from 1/100 to 1/800,
and ξ =0.9.
The Soft Bound Admission Control Algorithm 119
5 Conclusion
In this paper, we proposed SBAC algorithm that reduces the blocking probability of
vertical handover connections within the limited capacity of system over ubiquitous
environment (e.g. WLAN and WAAN). Proposed SBAC algorithm considers vertical
handover connections that have higher priority. In order to analyze the blocking prob-
ability of SBAC algorithm, we use the two-dimensional Markov chain model. From
the numerical analysis, we compared SBAC algorithm against traditional non-bound
and fixed bound scheme. As a result, proposed SBAC scheme is able to improve
handover blocking probability in ubiquitous environment. Future work needs to ana-
lyze the utilization based on optimized critical bandwidth ratio.
Acknowledgement. This work was supported in part by the MIC, Korea under the
ITRC program supervised by the IITA and the KOSEF under the ERC program.
References
1. Dusit Niyato, Ekram Hossain: Call Admission Control for QoS Provisioning in 4G Wire-
less Networks: Issues and Approaches, IEEE Network, Sept.-Oct. (2005) 5-11
2. I. Katzela, M. Naghshineh: Channel assignment schemes for cellular mobile telecommuni-
cation systems: A comprehensive survey, IEEE Personal Commun., vol. 3, June (1996)
10–31
3. Yuguang Fang, Yi Zhang: Call Admission Control Schemes and Performance Analysis in
Wireless Mobile Networks, IEEE Transactions on Vehicular Technology, Vol. 51, No. 2,
March (2002) 371-382
4. Emre A. Yavuz, Victor C. M. Leung: A Practical Method for Estimating Performance
Metrics of Call Admission Control Schemes in Wireless Mobile Networks, IEEE WCNC,
March (2005) 1254-1259
5. M. Buddhikot et al.: Design and Implementation of a WLAN/CDMA2000 Interworking
Architecture, IEEE Communications Magazine, Nov. (2003)
6. D. Johnson, C. Perkins and J. Arkko,: Mobility Support for IPv6, RFC 3775, June (2004)
7. Sheng-Tzong Cheng, Jian-Liang Lin: IPv6-Based Dynamic Coordinated Call Admission
Control Mechanism Over Integrated Wireless Networks, IEEE Journal on Selected Areas
in Communications, Vol. 23, No. 11, Nov. (2005) 2093-2103
The Soft Bound Admission Control Algorithm 121
8. Reininger D., Izmailov R.: Soft quality-of-service control for multimedia traffic on ATM
networks, Proceedings of IEEE ATM Workshop, (1998) 234-241
9. Sung H. Kim, Yeong M. Jang: Soft QoS-Based Vertical Handover Scheme for WLAN and
WCDMA Networks Using Dynamic Programming Approach, LNCS 2524, Nov. (2002)
707-716
10. Kleinrock. L.: Queueing System, Vol. 1 Theory, John Wiley and Sons, New York (1975)
11. Jiongkuan Hou, Yuguang Fang: Mobility-based call admission control schemes for wire-
less mobile networks, Wireless Communications and Mobile Computing. Wirel. Commun.
Mob. Comput. (2001) 1:269–282
ImprovingHandoffPerformancebyUsing
Distance-BasedDynamicHysteresisValue
1 Introduction
In cellular communications, handoff is the process of transferring the serving base
station (BS) of a mobile station (MS) from one to another when the MS moves across
a cell boundary. A call in progress could be forced to abort during handoff if suffi-
cient resources cannot be allocated in the new wireless cell. A properly designed
handoff algorithm is essential in reducing the switching load of the system while
maintaining the quality of service (QoS). In this paper hard handoffs in a cellular
system are concentrated on. The term handoff is henceforth used to refer to hard
handoff.
When an MS is traveling from its serving BS to the target adjacent BS, the prob-
ability of handoff is generally designed to maximize at the cell boundary. The
decision to initiate a handover may be based on different measurements [1]-[4] The
received signal strength (RSS) measurement is one of the most common criteria. Tra-
ditional handoff algorithms depend on comparing the differential signal power level
between the serving BS and target BSs to a fixed handoff hysteresis value h. This
hysteresis value is designed to reduce the ping-pong effect in the handoff procedure.
Therefore, selection of this hysteresis value becomes important for optimizing hand-
off performance. If h is too small, numerous unnecessary handoffs may be processed,
increasing the network burden. However, if h is too large, the long handoff delay may
result in a dropped-call or low QoS.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 122 – 131, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Improving Handoff Performance 123
Two important performance indicators of a handoff algorithm are the average num-
ber of handoffs and the average handoff delay, both of which are required to be mini-
mized. The size of the handoff area is a very important criterion relating to handoff.
The handoff area should be small enough to avoid cell-penetration or cell-dragging.
Cell-dragging occurs when excessive proportion of mobiles moves a considerable
distance into neighboring cell areas without making a handoff, resulting in an in-
creased level of system interference and a decrease in system throughput. The smaller
the handoff area, the more performance improves. The standard deviation of the hand-
off location is an indicator of the size of the handoff area.
In order to improve handoff performance, various adaptive handoff algorithms
were proposed [5]-[7]. MSs have recently been given the ability to continuously track
the mobile location through various radio location techniques [8]-[10]. In this study,
an adaptive handoff algorithm is developed by dynamically determining the hysteresis
value as a function of the distance between the MS and the serving BS. Since the
handoff hysteresis value is varied based on MS’s location, it can intelligently reduce
the probability of unnecessary handoffs and maintain the QoS.
The paper is organized as follows: First, the adaptive handoff algorithm is pro-
posed in Section 2. Then, Section 3 analyzes the proposed handoff algorithm. Nu-
merical results are presented to demonstrate the performance improvement compared
with handoff algorithms with fixed hysteresis values. Performance evaluation is pre-
sented in Section 4 by comparing key performance criteria. The effect of distance
error is discussed in Section 5. Finally, concluding remarks are presented in Section 6.
dc 4
h = max{20(1 − ( ) ), 0} , (1)
R
where dc is the distance between the MS and serving BS, and R is the cell radius.
Handoff is initiated only if the new BS’s signal strength is sufficiently stronger by
hysteresis value h, than that of the serving BS.
The coefficient is chosen to be 20, to enable the handoff algorithm to control un-
necessary handoff and react to the deep fading simultaneously, resulting in sudden
drop (20~30 dB) of RSS. The large exponent results in a decrease in the number of
handoffs at the cost of increasing handoff delay and probability of link degradation. In
124 H. Zhu and K. Kwak
this paper, 4 is the default exponent for the proposed algorithm, because it is assumed
that the probability of link degradation should not exceed 0.005.
As demonstrated in Equation (1), h decreases from 20 to 0 dB as the MS moves
away from the serving BS. By setting the above dynamic hysteresis value h, the num-
ber of unnecessary handoff is decreased because of a large h if the MS is near the
serving BS, and the MS is encouraged to hand over to adjacent cells because of a
small h if it is near the boundary of the current cell. In this way, handoff area is opti-
mized.
It is assumed that the RSS is affected by path loss as well as the shadowing effect. In
addition, Rayleigh fading exists. However, this is averaged out and can be ignored
because handoff algorithms cannot respond to short-term fading [11]-[16]. Therefore,
if the transmitted power of BS is normalized to be 0 dB, the signal strengths received
from the current BS and adjacent BSs, denoted Rc and Ri, are given by
Rc = − K log(d c ) + u (d c ) , (2)
where K represents the path loss factor, dc and di represent the distance from the cur-
rent BS and adjacent BSs respectively, and u(d) and vi(d) model the effect of shadow-
ing fading.
where Ri and Rc are the RSS of the adjacent BSs and serving BS at the MS’s location,
respectively, and h is the hysteresis value. If the handoff condition is satisfied, the call
is handed over to the adjacent cell with the largest Ri. Handoff will not occur unless
the RSS from an adjacent BS is greater than that from the serving BS by the hystere-
sis value h. Therefore, the handoff probability Pho is given as:
6
Pho = P{* [ Ri > Rc + h]} , (5)
i =1
instants (or at two different positions of the MS), is not included in the above equa-
tions. However, dcor must be considered when a moving MS is considered for per-
formance evaluation.
Handoff probability for MS under the cover of the BS is computed at different loca-
tions. If signal power is noise free, the hysteresis value can be set to 0 such that the
MS only performs handoff at the cell boundary; in this case, handoff probability is 1
at the cell boundary and 0 elsewhere. When noise is present, the handoff probability is
zero near the center of the cell, and increases as the MS moves away from the serving
BS till handoff occurs. A good handoff algorithm should have a handoff probability
map approaching the ideal case.
1
h=0 dB
h=4 dB
0.8 h=10 dB
adaptive h
handoff probability
0.6
0.4
0.2
0
0 500 1000 1500
location (m)
Assuming the distance between two neighboring BS is 2000m, the handoff prob-
ability of the proposed algorithm when the MS moves away from the serving BS to an
adjacent BS along a straight line, is compared with that of handoff algorithm with
fixed h=0,4,10 dB, shown in Fig. 1. The algorithm demonstrates superior perform-
ance, because the handoff probability of the algorithm is smallest when the distance is
less than 950m, and increases most rapidly from 0 to 1 in the handoff area near the
cell boundary. Therefore, the proposed algorithm has the smallest handoff area.
In contrast to the handoff probability discussed above, for the MS locating in a cell
geographically but communicating with the BS of another cell, another performance
metric is the handoff probability from the serving BS to the BS of the cell in which
the MS locates geographically, denoted by Pb. In Fig. 2, the MS locates in Cell 0, so
its serving BS should be BS0, in order to balance the system traffic and reduce inter-
ference. However, its serving BS is BS1. In this case, Pb is the probability that the MS
126 H. Zhu and K. Kwak
hands over from Cell 1 to Cell 0, i.e., from BS1 to BS0. For the ideal case, Pb is one
within the cell area and zero elsewhere, if the signal power is noise free. Therefore,
for an effective handoff algorithm, Pb should be one in most of the cell area and de-
crease rapidly in the area near the cell boundary.
Fig. 2. Illustration of Pb
Assuming the serving BS is BSj, j=1,2,…6, handoff occurs from BSj to BS0 if the
following conditions are met:
1) If the RSS from BS0 exceeds that of BSi for any ij;
2) If the RSS from BS0 exceeds that of BSj by hysteresis level h.
The corresponding handoff probability is
6
Pb j = P{ [ R c > Ri ] [ Rc > R j + h]} . (6)
i =1, i ≠ j
Therefore,
6
Pb = ¦ PBS j Pb j (7)
j =1
where PBSj is the probability that the MS’s serving BS is BSj, j=1,2,…,6.
Since PBSj is extremely difficult to compute, if not impossible, the assumption
that they have the same probability is taken, i.e., PBSj =1/6, j=1,2,…,6. Fig. 3 illus-
trates the comparison of Pb between handoff algorithms with adaptive and fixed
hysteresis when the MS moves away from the serving BS to an adjacent BS along
a straight line. The proposed algorithm demonstrates superior performance, because
the handoff probability of the proposed algorithm is the highest when the distance
is less than 900m, and decreases most rapidly in the handoff area near the cell
boundary.
Improving Handoff Performance 127
0.9
0.8
handoff probability
0.7
0.6
0.5
0.4 h=0 dB
h=4 dB
0.3
h=8 dB
0.2 h=10 dB
adaptive h
0.1
0 200 400 600 800 1000 1200
location (m)
Fig. 3. Comparison of Pb
4 Performance Evaluation
The parameters used for simulation are presented in Table 1, which are commonly
used to analyze handoff performance [11]-[15]. The MS is assumed to move from the
serving BS to an adjacent BS along a straight line. The average number of handoffs,
average handoff delay, standard deviation of handoff location, average signal strength
during handoff and probability of link degradation are used as criteria for perform-
ance evaluation.
4.1 Comparison with Handoff Algorithm Using Both Threshold and Hysteresis
In regards to the handoff algorithm using both threshold and hysteresis, i.e., the hand-
off algorithm using both absolute and relative signal strength [11], the MS is handed
over from one BS to another if both of the following conditions are met:
1) The RSS of the serving BS falls below an absolute threshold value T dB;
2) The RSS of the new BS becomes greater than that of the serving BS by a
hysteresis of h dB.
128 H. Zhu and K. Kwak
Fig. 4 illustrates the comparison of handoff probability for handoff algorithms with
adaptive hysteresis, fixed hysteresis, and both threshold and fixed hysteresis when the
MS moves along a straight line from the serving BS to an adjacent BS. The handoff
probability of the proposed algorithm is the smallest when the distance is less than
900m, and is the largest when the distance is greater than 1100m. The proposed algo-
rithm is superior because it can control handoff when the MS is near the serving BS,
and encourages handoff when the MS is leaving the serving cell.
1
h=4 dB, T=-85 dB
h=4 dB, T=-90 dB
0.8 h=4 dB, T=-95 dB
h=4 dB, T=-100 dB
handoff probability
fixed h=4 dB
0.6 adaptive h
0.4
0.2
0
0 500 1000 1500
location (m)
proposed adaptive handoff algorithm has smaller handoff delay and standard devia-
tion of handoff area, which will decrease the interference on the neighboring cell and
increase system throughput. Signal strength during handoff and the probability of link
degradation of the proposed algorithm are much better. The improvements of the
proposed algorithm are remarkable: 68.93% for handoff delay, 58.38% for standard
deviation of handoff location, 6.334dB for average signal strength during handoff,
and 70.07% for probability of link degradation.
As for the handoff algorithm based on distance and hysteresis [12], a handoff is per-
formed to the adjacent base station if both the following conditions are met:
1) If the RSS from the adjacent BS exceeds that of the serving BS by a hystere-
sis level h dB;
2) If the measured distance from the serving base station exceeds that of the ad-
jacent station by a threshold distance Ȗm.
Table 4 demonstrates that this handoff algorithm can effectively decrease the num-
ber of handoffs by using a large distance threshold. The average number of handoffs
decreases with increasing distance threshold at the cost of large handoff delay and
high probability of link degradation.
The probability of link degradation is extremely high for the handoff algorithm
based on distance and hysteresis, because the handoff algorithm cannot react to a sud-
den drop in signal strength when the MS is in the area where the distance condition of
handoff is not satisfied. The call is forced to interrupt even though there is another
feasible link, because handoff is not performed. This is also the reason why the stan-
dard deviation of handoff location is extremely small. From the subscriber’s point of
view, the blocking of handoff is less desirable than the blocking of a new call. There-
fore, the handoff algorithm based on distance and RSS is not feasible because of its
high probability of link degradation when call-drop probability is considered.
Table 5 presents the results for different levels of location accuracy. The average
number of handoffs increases with std, while average handoff delay decreases as std
increases. There is a little change for the standard deviation of handoff area, signal
strength during handoff and probability of link degradation. It can be concluded that
the adaptive handoff algorithm is stable when the standard deviation of the meas-
ured distance is less than 60m, meaning that the proposed algorithm is still feasible
when only rough distance information rather than accurate distance information is
available.
6 Conclusion
Acknowledgment
This research was supported by University IT Research Center Project of Inha UWB-
ITRC, Korea. The work of Huamin Zhu was supported in part by Korea Science and
Engineering Foundation (KOSEF).
Improving Handoff Performance 131
References
1. Tripathi, N.D., Reed, J.H., VanLandinoham, H.F.: Handoff in cellular systems, IEEE
Wireless Commun.unications, Vol. 5, No. 6 (1998) 26-37B
2. Pollini, G.P.: Trends in handover design, IEEE Communications Magazine, Vol. 34, No. 3
(1996) 82–90
3. Tekinay, S. and Jabbari, B.: Handover and channel assignment in mobile cellular net-
works, IEEE Commun. Mag. (1991) 42-46
4. Zonoozi, M., Dassanayake, P., Faulkner, M.: Optimum hysteresis level, signal averaging
time and handover delay, Proc. IEEE 47th Veh. Technol. Conf., Vol. 1 (1997) 310-313
5. Vijayan, R. and Holtzman, JM.: A model for analyzing handoff algorithms, IEEE Trans.
on Vehic. Techn., Vol. 42, No. 3 (1993) 351-356
6. Wang, S.S, Green, M. and Malkawi, M.: Adaptive handoff method using mobile location
information, Broadband Communications for the Internet Era Symposium digest, 2001
IEEE Emerging Technologies Symposium on (2001) 97-101
7. Lau, S.S.-F., Cheung, KF, Chuang, J.C.I: Fuzzy logic adaptive handoff algorithm, Global
Telecommunications Conference, 1995. GLOBECOM '95, IEEE, Vol. 1 (1995) 509-513
8. Fang, B.T.: Simple solutions for hyerbolic and related position fixes, IEEE Trans. Aerosp.
Electron. Syst., Vol. 26 (1990) 748-753
9. Wallehnhof, H., Lichtenegger, H., and Collins, J.: Global Positioning System: Theory and
Practice. Springer-Verlag, New York (1997)
10. Bajaj, R., Ranaweera, S.L., and Agrawal, D.P.: GPS: Location-Tracking technology, Com-
puter, Vol. 35 (2002) 92–94
11. Zhang, N. and Holtzman, JM.: Analysis of handoff algorithms using both absolute and
relative measurements, IEEE Trans. on Vehic. Techn., Vol. 45, No. 1 (1996) 174–179
12. Itoh, KI, Watanabe, S., Shih, J.S., and Sato, T.: Performance of handoff algorithm based
on distance and RSSI measurements, IEEE Trans. on Vehic. Techn., Vol. 51, No. 6 (2002)
1460-1468
13. Prakash, R. and Veeravalli, V.V.: Adaptive hard handoff algorithms, IEEE J. Select. Areas
Commun., Vol. 13, No. 11 (2000) 2456-2464
14. Veeravalli, V.V. and Kelly, O.E.: A locally optimal handoff algorithm for cellular commu-
nications, IEEE Trans. on Vehic. Techn., Vol. 46, No. 3 (1997) 351-356
15. Prakash, R. and Veeravalli, V.V.: Locally optimal soft handoff algorithms, IEEE Trans. on
Vehic. Techn., Vol. 52, No. 2 (2003) 347-356
16. Gudmundson, M.: Correlation model for shadow fading in mobile radio systems, Electron.
Lett, Vol. 27, No. 23 (1991) 2145-2146
A Seamless Service Management with Context-Aware
Handoff Scheme in Ubiquitous Computing Environment
1 Introduction
Recently, ubiquitous computing is regarded as one of the most attractive research
topics in computer science, and thus both academic and industrial sectors conduct
research in many relevant areas. One of the major premises for ubiquitous computing
is that all components should be connected to the network [1]. For this reason, seam-
less connectivity is an important research issue in ubiquitous computing. Also to pro-
vide seamless connectivity, a handoff management scheme between heterogeneous
network interfaces, called VHO, considering different characteristics of each interface
and various context informations is required.
However, existing approaches only assume that all components are connected with
networks seamlessly, and do not consider how seamless connectivity is provided.
Also in the aspect of seamless service, existing research only focuses on adaptation of
application itself or resource reservation to support quality of service (QoS). And they
only consider one network interface and pay no attention to VHO between heteroge-
neous network interfaces. Thus, if current network interface does not satisfy a given
condition, they cannot keep performing services. At the same time, in VHO schemes,
existing handoff operation is only focused on safe and reliable handoff, and it only
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 132 – 141, 2006.
© Springer-Verlag Berlin Heidelberg 2006
A Seamless Service Management with Context-Aware Handoff Scheme 133
occurs based on network factors, such as signal strength and latency without consider-
ing any context information, e.g., user intentions.
To provide seamless connectivity, this paper proposes an effective application exe-
cution model to provide seamless service by changing application service mode adap-
tively and proposes an intelligent context-aware handoff management scheme based
on the profiles reflecting various context informations in ubiquitous computing envi-
ronment composed of heterogeneous network interfaces. We define our own profiles
and then we design basic modules organizing a middleware architecture for seamless
service management and efficient VHO management, especially an application agent
that changes application’s service mode to control service quality for seamless ser-
vice. Consequently, proposed seamless service management with context-aware
handoff scheme considering user intentions as well as the issues of network layer
provides service continuity in ubiquitous computing environment by preventing un-
necessary handoffs, changing application service modes adaptively, and eventually
maintaining optimal network configuration. And also simulation results show about
38% enhanced performance in the processing rate, about 13% enhanced performance
from service continuity, and about 29% enhanced performance in the processing time,
compared with conventional schemes.
The remaining part of this paper consists of four parts. In Section 2, related work is
introduced. In Section 3, proposed schemes, middleware architecture, and algorithms,
are presented. Section 4 shows evaluation result and, lastly, conclusions are presented
in Section 5.
2 Related Work
In this section, precious research about ubiquitous computing, handoff schemes and
seamless service will be described.
Pervasive computing infrastructure, GAIA, allows applications to specify different
behaviors in different contexts by gathering of context information from different
sensors, delivering appropriate context information to the applications and providing a
powerful context model that allows complex reasoning to be done on contexts [2]. An
architectural framework, AURA, allows user mobility and adapts computing envi-
ronments proactively to a given situation by using models of tasks consisted of in-
ferred service information and user information [3]. Context Toolkit [4] presents
conceptual model for context-aware applications and enables the construction of pro-
totypes for context-aware applications. Another pervasive computing infrastructure
that exploits Semantic Web technologies to support explicit representation, expressive
querying, and flexible reasoning of contexts in smart spaces, Semantic Space, enables
applications to run proactively [5]. But, these approaches are about techniques for
ubiquitous computing, like sensor abstraction, context reasoning, and so on, based
on the assumption that seamless connectivity is guaranteed. They did not consider
any interaction mechanism between changing environment and network status [2] [3]
[4] [5].
In [6], they propose an adaptive QoS design approach via resource reservation and
rate adaptation to support multimedia over wireless cellular networks. In [7], they
propose the optimization of the QoS offered by real-time multimedia adaptive
134 T.-H. Kang et al.
To support seamless service and seamless connectivity, optimal application and net-
work configuration should be guaranteed. Thus, the proposed middleware architecture
A Seamless Service Management with Context-Aware Handoff Scheme 135
consists of application profile manager, user profile manager, context engine, applica-
tion manager, application agent, and VHO decision manager shown in Fig. 1.
Application profile manager manages application profiles reflecting some impor-
tant application requirements. Application profile consists of several service modes
that represent different service quality and minimum required values for bandwidth,
packet error rate (PER), and latency in each service mode as shown as in Table 1.
User profile manager is to manage user profiles reflecting any specific user informa-
tion and requirements. User profile is composed of using time, transferred data volume
and priority in each work. Using time is the average time for a given work process com-
pletion in the previous work process and transferred data volume is the average amount
of transferred data during the same process. Priority is the degree of preference for appli-
cations in each work. Example of user profile is shown as in Table 2.
Application Name Work Using Time (s) Transferred Data Volume (Kbytes) Priority
Navigation Navigation 1920 8192 1
Work Name
Navigation
Using Transferred Current Band- La-
Applica- Prior- PER
Time Data Volume Service Service Mode width tency
tion Name ity (%)
(s) (Kbytes) mode (kbps) (ms)
1(video-base) 192 5 10
Naviga-
1920 8192 1 1 2(audio-base) 64 8 12
tion
3(text-base) 16 10 15
Work profile manager manages work profile, requests application invocation or the
change of service mode, and notifies application suspension to the application layer.
136 T.-H. Kang et al.
Application manager generates abstract profile from the work profile. The abstract
profile consists of bandwidth, packet error rate, latency, and work bandwidth. Band-
width is the sum of all application’s required bandwidths. Packet error rate and la-
tency are determined by all application’s requirements, which may be the minimum
value among all application’s values. Work bandwidth represents a permitted band-
width based on user pattern and is used to prevent unnecessary handoff during VHO
decision procedure. Its example is shown in Table 4.
Work bandwidth is calculated as follows:
sum of all application' s Transferred Data Volume
Work bandwidth =
Using Time of Work
Work Name Bandwidth (kbps) PER (%) Latency (ms) Work bandwidth(kbps)
Navigation 160 5 8 120
VHO decision manager decides whether it should perform any handoff and requests
the network layer to change any chosen network interface if necessary, based on the
abstract profile and network information. In this paper, we assume that network layer
provides the information about other available network interfaces as well as current
network interface through control channel. Detailed VHO decision algorithm is de-
scribed as in Fig. 2.
check current network whether bandwidth, PER, latency of Abstract Profile are satisfied
if satisfied all factors then
stay in current network
else
check current network whether work bandwidth, PER, latency of Abstract Profile are satisfied
if satisfied all factors then
stay in current network
else
check available network whether bandwidth, PER, latency are satisfied
if satisfied network exist then
handoff to satisfied network
else
check available network whether work bandwidth, PER, latency are satisfied
if satisfied network exist then
handoff to satisfied network
else
handoff to nearest network to requirement
request Application Agent for Service mode change
Application agent changes service mode of any application to maintain service quality
for seamless service. When VHO decision manager requests any service mode to
be changed, application agent checks whether current network may satisfy work
bandwidth or not. If current network is satisfied for work bandwidth, then enhance all
A Seamless Service Management with Context-Aware Handoff Scheme 137
application’s service quality and reactivate any suspended application if exists, and
deliver this result to the context engine to update work profile. If current network does
not satisfy work bandwidth, application agent temporally enhance all application’s
service quality and reactivate any suspended application if exists. Then application
agent changes application’s service mode based on each application’s priority speci-
fied in work profile in the order of lower priority. If current situation is not improved
by changing service mode, then the application agent suspends any application in the
order of lower priority and delivers this result to the context engine to update work
profile. Detailed application’s service configuration procedure of application agent is
described as in Fig. 3.
Start
dissatisfy
satisfy
The overall operation progress based on the proposed scheme is performed as in the
following steps. In the first step, context engine gathers user/application profile to gen-
erate its work profile and delivers it to the work profile manager. In the second step,
work profile manager requests an application invocation to the application layer and
delivers its work profile to the application manager. In this step, work profile manager
requests to changes current service mode with suspension of any application to the ap-
plication layer and notifies this result to user. In the third step, application manager
generates abstract profile based on work profile and delivers it to the VHO decision
manager. In the fourth step, VHO decision manager decides whether it should perform
handoff and requests VHO to the network layer. And then application’s service mode is
138 T.-H. Kang et al.
4 Evaluation
In this section, we present simulation results to compare the proposed scheme with
some other VHO schemes, conventional VHO scheme that only concerns network
issues, and AAO (Active Application Oriented) scheme that concerns network issues
and applications requirements.
Simulation is performed with three network interfaces, seven applications, and four
work groups. The three network interfaces are WWAN, DMB, and Wibro, and network
factors about each interface are based on [10] [11] [12]. While simulation is proceeded,
network factors of these network interfaces, bandwidth, PER, and latency, are randomly
changed based on each interface's characteristics and the list of available network inter-
faces are randomly changed too. And theoretical specification of each network interface
is shown as in Table 5. Seven applications are chosen as navigation service, broadcast-
ing service, data transmission service, Instant Messaging Service (IMS), online game
service, explorer service, and music service. Also four work groups are shown as in
Table 6. In every simulation, applications for each work group have different using
times and transferred data volumes, which are randomly generated.
Wibro 1 Mb 1.5 20 1
DMB 15 Mb 2.5 10 5 ~ 30
Through the simulation result, three kinds of data can be obtained, i.e., the number of
VHOs, the number of application failures, and the throughput. And based on these
three kinds of data, we evaluate our proposed scheme compared with conventional
scheme and AAO schemes in terms of three aspects, i.e., processing rate, service
continuity and processing time.
Fig. 4 represents the processing rate, which is the percentage of overall throughput
when ideal network environment exists. And according to the simulation result, the
proposed scheme shows about 38% enhanced performance compared with conven-
tional scheme and AAO scheme.
120
100
80
Continuity (%)
60
40
20
0
6:05
6:50
7:20
7:57
8:15
8:45
9:20
10:00
10:31
11:10
11:55
12:30
13:15
13:40
14:15
14:40
15:25
16:00
16:45
17:30
18:04
18:40
19:15
19:54
Time (m)
Fig. 5, 6, 7 represents the service continuity, which shows the amount of continu-
ous computation calculated based on VHO delay and application’s status reflecting
any application suspension, namely application failure. In other words, service conti-
nuity represents how seamless service is maintained. And according to the simulation
result, the proposed scheme in Fig. 7 shows about 13% enhanced performance com-
pared with conventional scheme in Fig. 5 and AAO scheme in Fig. 6.
120 120
100 100
80 80
Continuity (%)
Continuity (%)
60 60
40 40
20 20
0 0
6:05
6:50
7:20
7:57
8:15
8:45
9:20
10:00
10:31
11:10
11:55
12:30
13:15
13:40
14:15
14:40
15:25
16:00
16:45
17:30
18:04
18:40
19:15
19:54
6:05
6:50
7:20
7:57
8:15
8:45
9:20
10:00
10:31
11:10
11:55
12:30
13:15
13:40
14:15
14:40
15:25
16:00
16:45
17:30
18:04
18:40
19:15
19:54
Fig. 6. Service continuity of AAO scheme Fig. 7. Service continuity of proposed scheme
140 T.-H. Kang et al.
Fig. 8 represents the processing time, which is the time required for processing any
specific amount of task. In Fig. 8, processing time of work group-1 shows the time
spent for processing 450Mb task, processing time of work group-2 is the time spent
for processing 600Mb task, processing time of work group-3 is the time spent for
processing 450Mb task, processing time of work group-4 is the time spent for proc-
essing 450Mb task. And according to the simulation result, the proposed scheme
shows about 29% enhanced performance compared with conventional scheme and
AAO schemes.
Processing time
160
140
Processing time (m)
120
100
80
60
40
20
0
1 2 3 4
Work
5 Conclusion
Seamless connectivity is an important research issue in ubiquitous computing. How-
ever, research for seamless connectivity has not been considered properly and existing
approaches only assume that all components are connected with networks seamlessly
or are only concerned about factors of network issues. The provision of seamless
connectivity requires an intelligent handoff management which builds up the optimal
network environment and an application execution model which guarantees seamless
service by configuring service quality of applications adaptively.
In this paper, we propose a context-aware handoff management scheme for seam-
less connectivity and a seamless service management scheme for continuous service
in ubiquitous computing environment, based on profiles reflecting various context
informations, such as user intentions as well as the issues of network layer. Conse-
quently, the proposed schemes prevent unnecessary handoff operations, provide
service continuity, and eventually enable an optimal network configuration in ubiqui-
tous computing environment. And also simulation results show about 38% enhanced
performance in the processing rate, about 13% enhanced performance from service
continuity, and about 29% enhanced performance in the processing time, compared
with conventional schemes.
A Seamless Service Management with Context-Aware Handoff Scheme 141
References
1. Weiser, M.: The computer for the 21st century. Scientific American, vol. 265 (3), Septem-
ber 1991, page(s) 94-104.
2. Ranganathan, A. and Campbell, R. H.: An infrastructure for context-awareness based on
first order logic. Personal Ubiquitous Computing, vol. 7 (6), 2003, page(s) 353-364.
3. Sousa, J. and Garlan, D.: Aura: an architectural framework for user mobility in ubiquitous
computing environments. Proceedings of the 3rd Working IEEE/IFIP Conference on Soft-
ware Architecture, August 2002, page(s) 29-43.
4. Dey, A.K., Salber, D. and Abowd, G.D.: A conceptual framework and a toolkit for sup-
porting the rapid prototyping of context-aware applications. Human-Computer Interaction
(HCI) Journal, Vol. 16 (2-4), 2001, page(s) 97-166.
5. Wang, X., Dong, J.S., Chin, C.Y., Hettiarachchi, S.R., and Zhang, D.: Semantic Space: an
infrastructure for smart spaces. IEEE Pervasive Computing, Vol. 3 (3), July-September
2004, page(s) 32-39.
6. Lu, S.W., Lee, K.W. and Bharghavan, V.: Adaptive Quality of Service Support for Packet-
Switched Wireless Cellular Networks. Multimedia Tools and Applications, Vol. 17 (2-3),
July 2002, page(s) 157–179.
7. Ruiz, P.M., Botia, J.A. and Gomez-Skarmeta, A.,: Providing QoS through machine-
learning-driven adaptive multimedia applications. IEEE Transactions on Systems, Man
and Cybernetics, Part B, Vol. 34 (3), June 2004, page(s) 1398-1411.
8. Chen, W.T. and Shu, Y.Y.: Active Application Oriented Vertical Handoff in next-
Generation Wireless Networks. IEEE Wireless and Communications and Networking Con-
ference 2005, Vol. 3(13-17), March 2005, page(s) 1383-1388
9. Inoue, M., Mahmud, K., Murakami, H., Hasegawa, M. and Morikawa, H.: Context-Based
Network and Application Management on Seamless Networking Platform. Wireless Per-
sonal Communications, Vol. 35 (1-2), October 2005, page(s) 53-70.
10. Telecommunications Technology Association (TTA), http://www.tta.or.kr/
11. GSM World, GPRS-Standard class 10, http://www.gsmworld.com/
12. Media Independent Handover (MIH) Working Group, http://www.ieee802.org/21/
Performance Analysis of an Adaptive Soft Handoff
Algorithm for Mobile Cellular Systems
1 Introduction
Handoff is an essential component of mobile cellular communication systems [1] [2].
It is the process whereby a mobile station (MS) communicating with one base station
(BS) is switched to another BS when the MS moves across a cell boundary during a
call. A call in progress could be forced to abort during handoff if sufficient resources
cannot be allocated in the new wireless cell. A properly designed handoff algorithm is
essential in reducing the switching load of the system while maintaining the quality of
service (QoS). The design of reliable handoff algorithms is crucial to the operation of
a cellular communication system and is especially important in microcellular systems,
where the MS may traverse several cells during a call. The decision to initiate a hand-
off may be based on different measurements, such as the received signal strength
(RSS) from the serving BS and neighboring BSs, the distance between the MS and the
surrounding BSs, signal to noise ration (SNR), and bit error rate (BER). The RSS
measurement is one of the most common criteria.
There are two types of handoff: hard handoff and soft handoff. Hard handoff is a
break-before-make method, where a new link is set up after the release of the old link.
A certain amount of margin is introduced to eliminate the ping-pong effect, which is
the scenario of repeated handoff between two adjacent BSs caused by rapid fluctua-
tions in the RSS from both of the BSs. Soft handoff is a make-before-break method
[3]-[6]. With soft handoff, an active set is maintained, which is the set of all the BSs
with which an MS is communicating. Depending on the changes in RSS from the two
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 142 – 151, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Performance Analysis of an Adaptive Soft Handoff Algorithm 143
or more BSs involved, a hard decision will eventually be made to communicate with
only one BS. This normally happens after it is clear that the signal from one BS is
considerably stronger than those from the others. In the interim period, the MS has
simultaneous communication with all BSs in the active set. Generally, three parame-
ters are specified in a soft handoff algorithm: the add threshold Tadd, the drop thresh-
old Tdrop, and the drop timer Dtime. When the pilot signal from a new BS exceeds that
from the serving BS by a threshold value Tadd, a new link to the new BS is established
while maintaining the existing link. In this case, the call is said to be in soft handoff.
We here assume that an MS can be in soft handoff with two strong BSs. When the
RSS from either the old BS or the new BS weakens below Tdrop and remains there for
Dtime, the bad connection is released and only a single good connection is maintained.
The MS should reset and disable the timer if the level of RSS goes above the drop
threshold Tdrop before the timer expires.
The rest of this paper is structured as follows. The proposed adaptive soft handoff
algorithm is presented in Section 2. Performance metrics to measure the performance
of soft handoff are derived in Section 3. In Section 4, the simulation environment is
described and the proposed adaptive handoff algorithm is compared with the static
soft handoff algorithm. Finally, concluding remarks are presented in Section 5.
encourage other BSs in the active set to leave the active set when the RSS from a
serving BS is much higher than the link degradation threshold.
A simple adaptive soft handoff algorithm with dynamic threshold values is pro-
posed, which is determined by the RSS from the primary BS in the active set, i.e.
where the primary BS is the serving BS with the largest RSS in the active set, sp is the
RSS from the primary BS, is link degradation threshold.
Therefore, the following equation can be derived:
°10, s p > Δ + 15
Tadd = ® . (3)
°̄s p − Δ − 15, s p ≤ Δ + 15
As demonstrated above, Tadd equals 10 dB for sp larger than +15, and it de-
creases from 10 dB to -5 dB as the RSS from the primary BS decreasing from +15
to . Number 15 is chosen because the link degradation probability at next time
instant, denoted Pld, is considerably low, if the MS is served by the same BS only. In
this case, Pld is approximately equal to Q(15/ı), where Q( ) is the Q-function (com-
plementary cumulative distribution function), ı is the standard variance of the shadow
fading process. For instance, the link degradation probability is about 0.0062 for ı=6.
Since the handoff thresholds are varied based on RSS, it can intelligently reduce
the probability of unnecessary handoffs while maintaining the signal quality. The
handoff performance criterion is based on minimizing both the average number of
active set updates and the mean size of the active set for the given propagation pa-
rameters and mobile path. The tradeoff between these two conflicting criteria is exam-
ined by comparing the proposed algorithm with a handoff algorithm with static
thresholds in Section 4.
3 Performance Metrics
For the sake of simplicity, a basic system consisting of two BSs separated by a dis-
tance of D is considered in this paper [2]-[5]. And it is assumed that the MS is moving
along a straight line with a constant velocity v between the two BSs, labeled BS1 and
BS2. It is assumed that the RSS is affected by path loss as well as the shadowing ef-
fect. In addition, Rayleigh fading exists. However, this is averaged out and can be
ignored because handoff algorithms cannot respond to short-term fading [3]-[6].
Let di(n) denote the distance between the MS and BSi, i=1,2 at time instant n.
Therefore, if the transmitted power of BS is normalized to be 0 dB, the signal strength
from BSi, denoted si(n), i=1,2, can be written as
where K represents the path loss factor, and ui(n), i=1,2 are zero mean Gaussian ran-
dom processes that model the log-normal shadow fading. The log-normal shadowing
is assumed to have an exponential autocorrelation function [7] [8]
where ı is the standard variance of the shadow fading process, and a is the correlation
coefficient of the discrete-time fading process.
a = e − vts / dc (6)
where ts is the sampling time and dc is the correlation distance determining how fast
the correlation decays with distance.
The difference between si(n) is defined as
Let P1ĺ12(n), P2ĺ12(n), P12ĺ1(n), and P12ĺ2(n), be the probabilities that BS2 is
added, BS1 is added, BS2 is dropped, and BS1 is dropped at time instant n, respec-
tively. These probabilities can be computed as the following:
{
P12→1 (n) = Pr − x(n − M ) < Tdrop − x(n − M − 1) ≥ Tdrop }
{ }
n (10)
⋅ ∏ Pr − x(k ) < Tdrop − x(k − 1) < Tdrop
k = n − M +1
{
P12→2 (n) = Pr x(n − M ) < Tdrop x(n − M − 1) ≥ Tdrop }
∏ Pr{x(k ) < T }
n (11)
⋅ drop x(k − 1) < Tdrop
k = n − M +1
where
M = ¬Dtime / t s ¼ (12)
Let us define P1(n), P2(n), and P12(n) as the probabilities that the active set contains
BS1 only, BS2 only, or both BS1 and BS2. Once the above transition probabilities are
found, the assignment probabilities P1(n), P2(n), and P12(n) can be calculated under
the initial condition P1(0)=1, P2(0)=0, P12(0)=0.
Two performance measures, average number of BSs in the active set NOBS (i.e.,
average size of the active set) and average number of active set updates NOupdate, are
given by
146 H. Zhu and K. Kwak
N
1 (13)
NOBS = ¦ [ P ( n ) +P ( n ) + 2 P
1 2 12 (n)]
N n =1
N
NOupdate = ¦ {P1 (n − 1) P1→12 (n) +P2 (n − 1) P2→12 (n)
n =1 . (14)
+ P12 (n − 1)[ P12→1 (n) + P12→2 (n)]}
However, the analytical results for the proposed adaptive soft handoff are intracta-
ble, because the thresholds Tadd and Tdrop are changing with the RSS. We will resort to
simulation for performance evaluation in the next section.
4 Performance Evaluation
The parameters used for simulation are presented in Table 1, which are commonly
used to analyze handoff performance [3]-[6]. The MS is assumed to move from the
serving BS BS1 to an adjacent BS BS2 along a straight line. The average number of
active set updates, the mean size of the active set, the mean time of soft handoff, the
average signal quality and link degradation probability are used as criteria for per-
formance evaluation. 50000 realizations were used to estimate the performance at
each parameter setting in order to get stable results and smooth curves.
The tradeoff between the average number of active set updates and the mean size
of the active set for soft handoff algorithms with different threshold values is illus-
trated in Fig. 1. Tadd = -1dB and Tdrop = -3dB are denoted as (-1,-3) and so on. It is
clear that the proposed adaptive soft handoff algorithm has better performance than
those with static thresholds. For soft handoff algorithm with static thresholds, the
average number of active set updates NOupdate decreases with the decrease of the drop
threshold at the cost of increased mean size of active set NOBS for the same add
threshold; On the other hand, as the add threshold increases, both NOupdate and NOBS
decrease for the same drop threshold. The cost accompanied with large add threshold
is the increasing link degradation probability, shown in Fig. 2.
Parameters Description
D=2000 m Distance between two BSs
R=D/sqrt(3) Cell radius
K=30 Path loss factor
ı=8 dB Shadow fading standard deviation
dc=30 m Correlation distance
v=20 m/s Mobile velocity
ts=0.5s Sample time
M=3 Drop timer 1.5s
ǻ= -105 dB Threshold of link degradation
Performance Analysis of an Adaptive Soft Handoff Algorithm 147
18
static
14 (-1,-5)
(-3,-7)
13 (-5,-9)
(0,-5)
12
(-1,-7)
11 (-3,-9)
(0,-7) (-5,-11)
10
1.1 1.15 1.2 1.25 1.3 1.35 1.4 1.45 1.5
mean size of active set
Fig. 1. Average number of active set updates versus average size of the active set
0.026
static threshold, Tdrop=-7
0.025
adaptive threshold
0.024 benchmark
link degradatio probability
0.023
0.022
0.021
0.02
0.019
0.018
0.017
-5 -4 -3 -2 -1 0
Tadd (dB)
Fig. 2 shows the effect of the add threshold when the drop threshold is fixed, and
Fig. 3 shows the effect of the drop threshold when the add threshold is fixed. The
benchmark is also shown in the figures as the extreme case where the MS is commu-
nicating with both BSs at all time. The link degradation probability increases with the
increase of either the add threshold or the drop threshold. The performance of the
proposed adaptive handoff algorithm is approaching that of the benchmark, and is
better than the handoff algorithm with static threshold values except the case with
(-5, -7) whose link degradation probability is a litter lower. However, the average
number of active set updates and mean size of active set of the handoff algorithm with
(-5, -7) is much worse than those of the proposed adaptive handoff, shown in Fig. 1.
148 H. Zhu and K. Kwak
0.032
static threshold, Tadd=0
0.03
adaptive threshold
benchmark
link degradatio probability 0.028
0.026
0.024
0.022
0.02
0.018
-7 -6 -5 -4 -3 -2 -1
Tdrop (dB)
0.9
0.8
0.7
0.6
probability
0.5
P1, (0,-5)
0.4 P1, (-1,-7)
P1,adaptive
0.3
P2, (0,-5)
0.2
P2, (-1,-7)
0.1 P2, adaptive
0
0 200 400 600 800 1000 1200 1400 1600 1800 2000
distance (m)
Fig. 4 shows the probabilities of assigning the MS to BS1 or BS2, and Fig. 5 shows
the probabilities of soft handoff, i.e., the MS is assigned to both BS1 and BS2. The
probability of soft handoff of the proposed adaptive algorithm is remarkably lower
than those of static algorithms, so we can expect that the mean time of soft handoff of
the proposed algorithm is shorter, shown in Table 2.
Comparison of link degradation probabilities at different location is shown in
Fig. 6. The proposed adaptive algorithm shows the best performance. The tradeoff
in decreasing the probability of soft handoff and enhancing link degradation
Performance Analysis of an Adaptive Soft Handoff Algorithm 149
perfor-mance is a minor decrease in signal strength. Three call quality curves are
compared in Fig. 7. The average signal strength decrease of the proposed algorithm is
around 0.4dB.
0.8
static (0,-5)
0.7 static (-1,-7)
adaptive
0.6
0.5
probability
0.4
0.3
0.2
0.1
0
0 500 1000 1500 2000
distance (m)
0
0 500 1000 1500 2000
distance (m)
0
static threshold (0,-5)
-10 static threshold (-1,-7)
adaptive threshold
-20
signal strength (dB)
-30
-40
-50
-60
-70
-80
-90
0 500 1000 1500 2000
distance (m)
5 Conclusion
An adaptive soft handoff algorithm with dynamic thresholds based on RSS is pro-
posed in this study. The proposed adaptive soft handoff algorithm can achieve smaller
average number of active set updates, reduce the mean size of the active set, and
lower the link degradation probability at the cost of a minor decrease of the signal
quality. Under the range of the parameters we have considered, the proposed adaptive
handoff algorithm significantly outperforms handoff algorithms with static thresholds.
Performance Analysis of an Adaptive Soft Handoff Algorithm 151
The only overhead of the proposed algorithm is calculating the thresholds according
to Eq. (1) and (2), which is extremely simple.
Acknowledgment
This research was partly supported by University IT Research Center Project of Inha
UWB-ITRC, Korea. The work of Huamin Zhu was supported in part by Korea Sci-
ence and Engineering Foundation (KOSEF).
References
1. Tripathi, N.D., Reed, J.H., VanLandinoham, H.F.: Handoff in cellular systems. IEEE Wire-
less Commun.unications, Vol. 5, No. 6 (1998) 26-37
2. Wong, D., Teng Joon Lim: Soft handoffs in CDMA mobile systems. IEEE Wireless Com-
munications, Vol. 4, No. 6 (1997) 6-17
3. Ning Zhang, Holtzman, J.M.: Analysis of a CDMA soft handoff algorithm. IEEE Transac-
tions on Vehicular Technology, Vol. 47, No. 2 (1998) 710-714
4. Wang, S.S., Sridhar, S., Green, M.; Adaptive soft handoff method using mobile location
information. IEEE 55th Vehicular Technology Conference, Vol. 4 (2002) 1936-1940
5. Akar, M., Mitra, U.: Soft handoff algorithms for CDMA cellular networks. IEEE Transac-
tions on Wireless Communications, Vol. 2, No. 6 (2003) 1259-1274
6. Prakash, R., Veeravalli, V.V.: Locally optimal soft handoff algorithms. IEEE Transactions
on Vehicular Technology, Vol. 52, No. 2 (2003) 347-356
7. Gudmundson, M.: Correlation model for shadow fading in mobile radio systems. Electron-
ics Letters, Vol. 27, No. 23 (1991) 2145-2146
8. Graziosi, F., Santucci, F.: A general correlation model for shadow fading in mobile radio
systems. IEEE Communications Letters, Vol. 6, No. 3 (2002) 102-104
An Admission Control and Traffic Engineering
Model for Diffserv-MPLS Networks
Haci A. Mantar1,2
1
Department of Computer Engineering, Gebze Institute of Technology, Turkey
2
Department of Computer Engineering, Harran University, Turkey
1 Introduction
A significant research effort has been done to support Quality of Services (QoS)
in the Internet. The research community has focused on three common archi-
tectures: Integrated Services (Intserv), Differentiated Services (Diffserv) [4] and
MultiProtocol Label Switching [3]. While Intserv with RSVP signaling provide
excellent QoS guarantees, it has scalability problems in the network core because
of the per-flow state maintenance and the per-flow operation in routers. Because
of scalability problem with this model, the IETF has proposed Diffserv [4] as
an alternative QoS architecture for the data/forwarding plane. Diffserv does not
have per-flow admission control or signaling and, consequently, routers do not
maintain any per-flow state or operation. Core routers merely keep states for
a small number of classes named Per Hop Behavior(PHB), each of which has
particular scheduling and buffering mechanisms. A packet’s PHB is identified
with the Diffserv code point field (DSCP) assigned by the ingress router.
To this end, Diffserv is relatively scalable with large network sizes because the
number of states in core routers are independent of the network size. Thus, it
is considered as the de facto QoS standard for the next generation of the Inter-
net. However, unlike the Intserv/RSVP, Diffserv only addresses data/forwarding
plane functionality, whereas control plane functions still remain open issues. 1)
A PHB defines the forwarding behavior in a single node. There is no QoS com-
mitment for the traffic traversing multiple nodes; 2) With the exception of Ex-
pedited Forwarding (EF), all the PHBs provide qualitative QoS guarantees (the
QoS metrics value changes with network conditions). Hence, the requirements of
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 152–161, 2006.
c Springer-Verlag Berlin Heidelberg 2006
An Admission Control and Traffic Engineering Model 153
We assume that a set of PHBs providing quantitative QoS guarantees are sup-
ported by each router [11]. A quantitative PHB i is associated with an upper
delay bound di , an upper loss ratio bound li , and certain percentage of link ca-
pacity Ci (e.g., di < 3ms, li < 0.01%). Each PHB can use only its share of link
capacity, and the surplus capacity of a PHB can be used by best-effort or quali-
tative services. We also assume that di and li are pre-determined at the network
configuration stage [8][9][11], which is done in relatively long time intervals (e.g.,
weeks) and downloaded into routers. A router dynamically adjusts its scheduler
rate and buffer size according to the dynamic traffic rate to meet pre-determined
di and li constraints. Under this premise, each router provides the desired QoS
regardless of the utilization rate. (For details on quantitative PHB, the reader
is referred to [8][9][11]).
To extend Diffserv QoS from a single node to across a domain, the IETF has
introduced Per-Domain Behavior (PDB)[5]. A PDB defines the upper bounds of
QoS constraints that identifiable packets will receive across a domain, regardless
of the network utilization, from ingress router to egress router (IR-ER). In this
sense, PDBs associates with QoS characteristics of LSPs.
Note that since the link resources allocated for a PHB can only be used by
that PHB, the network can be considered as if it is divided into multiple virtual
154 H.A. Mantar
networks, one for each PHB. Thus, in the rest of this paper, it is assumed that
there is only one PHB within a network.
For admission control and LSPs’ resizing, we use a BB [1]. Upon receiving a
reservation request, the proposed BB first determines the corresponding IR and
ER. It then checks the total resource availability in the associated LSPs (with
that particular IR-ER pair). If there is enough resource, it grants the requests,
otherwise it rejects.
As depicted in Figure 1, each router within the domain sends its link state
QoS information (for each interface-PHB pair) directly to the BB rather than
flooding to all other routers as done in traditional link state protocol. The ingress
routers also send the state of their LSPs (i.e., current traffic rate) to the BB.
The BB dynamically checks the traffic rate of LSPs. When the traffic rate of an
LSP reaches its maximum capacity, the BB attempts to resize the LSPs. The
new sizes of LSPs are determined according to their current utilization rate.
By resizing LSP with respect to their current utilization rate, overall network
load is balanced, congestion is minimized and admission control probability is
increased.
Once requests are accepted (by the BB), it is the role of ingress router to
distribute the aggregated load across LSPs according to their perceived path
costs, which reflects their utilization/congestion level. This is done in two phases,
compute the cost of LSPs and balance the load across them. The routers along
LSPs send their link cost to the associated ingress router periodically (e.g.,
every 5 sec). By having the cost of all the links along its LSPs, the ingress
router computes the cost of each LSP by simply adding the cost of all the links
constituting the LSP. It then attempts to equalize the cost of LSPs. The intuition
here is that if utilization of an LSP i is greater than utilization of an LSP j, then
the cost difference is minimized by shifting some load from i to j. This increases
the cost of j and decreases the cost of i and equilibrium state is reached when
they are equal. An important point here is that shifting does not violate QoS
commitments (Sec. 3.3), because the shifting process is done among the LSPs
that have the same QoS characteristics (the same QoS bound/PDB).
Each link in a domain has the cost as one of its QoS attributes. The idea behind
using the link and LSP cost is to shift the traffic from the congested path(s)
to the less congested one(s). Thus, the cost of a path should reflect the path’s
156 H.A. Mantar
congestion. Among many possible cost functions that exhibit this condition, we
choose a simple one. In this function a link cost for a PHB x is simply computed
as
cx (t) = qx /(1 − ux (t)) (1)
where qx is the fixed cost of using the link for x when it is idle, and ux (t)
represents the link utilization of x at time t. ux (t) = rx (t)/Rx , where rx (t) is
the traffic rate of x at time t and Rx represents the link capacity assigned to x.
The idea of the LSP cost update is very simple. Let L1, L2, L3, ..LN be the
links on an LSP i. The ingress router periodically receives the cost each of these
links. Once the ingress router has the cost of all links, it computes the cost of i
as
N
Ci = cj (2)
j=1
Given the traffic load flowing between an ingress-egress pair, the task of an
ingress router is to distribute the traffic across the candidate LSPs in a way
the loads are balanced and therefore congestion is minimized. This is achieved
through the cost-equalization of the LSPs. As described above, the ingress router
periodically computes its LSPs’ costs. When a persistent cost change is detected,
the ingress router invokes the load balancing/cost-equalization algorithm that
proportionally distributes the load across LSPs with respect to their perceived
cost. The idea here is that upon detecting a consistent and substantial cost
difference among LSPs, the ingress router shifts some of the traffic from the
high-cost LSP(s) to the low-cost LSP(s). The cost equalization can be done
using stochastic approximation theory or gradient projection methods, which
are generally complex to implement in practice. To circumvent this problem, we
use a simple but effective iterative procedure.
Let C1 , C2 , ...., CK be the costs, R1 , R2 , ...., RK the traffic rate and α1 , α2 , .....,
αK load proportions of LSP1 , LSP2 , ....., LSPK between an IR-ER pair and let
V be total load of a IR-ER pair. If all Ci are equal, then αi s are the desired
proportions. If not, we use mean costs of all the paths C − = Ci /K as the
target cost for each path and obtain new proportions. The new proportions are
computed as
C−
αi = αi (3)
Ci
1
To normalize αi , we define a normalization factor φ = K
αi
. The new values
i=1
are obtained as αi = φαi . The corresponding Ri ’s are: Ri = αi V . This procedure
is repeated iteratively until the costs of LSPs are equal. Because for a constant
LSP size (during the load balancing) the cost Ci is increased with its load (αi V ),
it can be shown that this procedure will always converge.
An Admission Control and Traffic Engineering Model 157
For clarification, consider Figure 1, where there are four LSPs, LSP1, LSP2,
LSP3, and LSP4. Let C1, C2, C3, and C4 be the costs and R1, R2, R3 and R4
be the traffic rate of LSP1, LSP2, LSP3 and LSP4, respectively. Suppose that
the ingress routers IR1 and IR2 detect substantial cost differences between their
LSPs, C2 > C1 and C3 > C4. In this situation, both IR1 and IR2 start shifting
some traffic from LSP2 to LSP1, and from LSP3 to LSP4, respectively. That
is, R1, R2, R3 and R4 will become R1+r1, R2-r1, R3-r2, and R4+r2,
respectively. This process is repeated until there is no appreciable cost difference.
Another important issue is how to assign the load to the LSPs. We use a
flow-based hashing approach in order to avoid out-of-order packet delivery. In
this approach, hashing is applied on source and destination IP addresses and
possibly other fields of the IP header. In the hashing model, the traffic will be
first distributed into N bins by using module-N operation on the hash space [6].
The router performs a modulo-N hash over the packet header fields that identify
a flow. If the total traffic rate is X bps, each bin approximately receives the
amount of X/N bps. The next step is to map N bins to LSPs. The number of
bins assigned to an LSP is determined based on its load portion.
The vc is computed for each link using equation (4). Each LSP has its own
share of a link as if it is the only one that uses that link. As seen in (4), when
several LSPs compute for the same resources, the resources are assigned based
on their utilization rate. After computing the vc of each link, the path size is
determined as
158 H.A. Mantar
4 Evaluation Results
In this section, we present experimental results to verify that the proposed model
is stable, robust and scalable in such a way that it minimizes congestion and
quickly balances the load cross multiple paths between IR-ER pairs in a reason-
able period of time. We modified our previous software tool [11] for this purpose.
We had a network topology consisting 8 edge routers (ingress, egress) and 8 core
routers (CR) (Figure 2). Edge routers had mesh connections. For each IR-ER
pair, two LSP were established, so there were 32 LSPs. Although an IR had LSP
connection to all the ERs, in the figure the connection to a single ER is shown
(for simplicity illustration).
Figure 3 and 4 shows, the load-balancing (cost-equalization) process for cost
update interval of 5 seconds and 1 seconds. In this experiment, the offered average
traffic rate between each IR-ER pair was changed every 60 sec. The figures shows
utilization of L1, L2 and L3 under different traffic conditions. As figures shows,
our model achieves load-balancing in few iterations.
Figure 5 shows the importance of our measurement-based paradigm in achiev-
ing the pre-determined QoS constraints. We verified QoS assurance only through
An Admission Control and Traffic Engineering Model 159
L1
IR1 CR1 CR5 ER1
L2
IR2 CR2 CR6 ER2
L3
IR3 CR7 ER3
CR3
90
70 L1 L2 L3
L1 L2 L3
80
60
70
50
Utilization Ratio(%)
60
Utilization Ratio(%)
40 50
30 40
30
20
20
10
10
0 0
5 30 55 80 105 130 155 180 205 230 255 280 1 11 21 31 41 51 61 71 81 91
Time(Seconds) Time(Seconds)
Fig. 3. The LSPs’ load balancing Fig. 4. The LSPs’ load balancing
loss ratio rate. For measurement intervals of 10 seconds and tolerable loss ra-
tio rate of 0.1%, the priority service met its given loss-ratio constraints with
99.92%. These QoS quantitative achievements come with the strict admission
control mechanism.
Figure 6 shows the comparison of the pre-established path scheme with the
shortest path and on demand path schemes. In on demand models such as [10],
the paths are computed based on requests, and the sessions are not rerouted,
meaning that there is no shifting process. As expected, the proposed model is
much better than the shortest path scheme when the load is high. The interesting
result here is that our scheme is even better than on demand scheme (it accepts
more requests). This is because of the lack of traffic shifting with the on demand
scheme. Note that the on demand path schemes also have serious scalability
problems. Because they performs the QoS routing and path reservation set up
for each individual requests.
As described before, upon receiving a reservation request, the BB performs
the admission control based only on the current utilization of the corresponding
LSPs. It does not check the capacity of the links constituting the LSPs. In on
demand scheme (the traditional approach of a BB), the BB first determines the
path and then check the capacity of all the links along the path. As shown in
Figure 7 and 8, our approach significantly decreases the BB admission control
time.
160 H.A. Mantar
Fig. 7. The BB admission control time Fig. 8. The BB pipe resizing frequency
Figure 7 depicts the BB pipe resizing frequency under different network load.
As shown, when the network is lightly-loaded, the resizing frequency is very
small. However, when the network is heavily-loaded (e.g., more than 90%),
the pipe resizing increases exponentially. This can result in serious scalability
problem, because the resizing process is computationally intensive. To reduce
this problem, the BB uses periodic update scheme for heavily-loaded network
conditions. For example, when the pipe resizing frequency exceeds some pre-
determined threshold, the BB resizes the pipes periodically. As shown 8, the
periodic update keeps the the resizing frequency in a reasonable range.
control time and minimizes scalability problems presented in prior research while
optimizing network resource utilization.
In this work we focused only on the quantitative services, in the future, we will
extend our model for the qualitative services. We will also study the theoretical
verification of stability.
References
1. K. Nichols, V. Jacobson, and L. Zhang. “ A Two-bit Differentiated Services Archi-
tecture for the Internet” RFC 2638, July 1999.
2. A. Elwalid, C. Jin, S. Low, and I. Widjaja, ”MATE: MPLS Adaptive Traffic En-
gineering”, IEEE INFOCOM 2001
3. E. Rosen, A. Viswanathan, R. Callon “Multiprotocol Label Switching Architec-
ture”, RFC 3031.
4. S. Black et al., “An Architecture for Differentiated Services,” RFC2475, Dec. 1998.
5. K. Nichols, B. Carpenter “Definition of Differentiated Services Per Domain Behav-
iors and Rules for their Specification” RFC 3086.
6. C. Hopps ”Analysis of an Equal-Cost Multi-Path Algorithm” RFC 2992, November
2000
7. S. Nelakuditi, Z.-L. Zhang, R.P. Tsang, and D.H.C. Du, Adaptive Proportional
Routing: A Localized QoS Routing Approach, IEEE/ACM Transactions on Net-
working, December 2002.
8. N. Christin, J. Liebeherr and T. Abdelzaher. A Quantitative Assured Forwarding
Service. In Proceedings of IEEE INFOCOM 2002.
9. S. Wang, D. Xuan, R. Bettati, and W. Zhao, ”Providing Absolute Differentiated
Services for Real-Time Applications in Static-Priority Scheduling Networks,” in
IEEE/ACM Transactions on Networking, Vol. 12, No. 2, pp. 326-339, April 2004.
10. P. Aukia, M.Kodialam,P.Koppol, “RATES: A server for MPLS Traffic Engineer-
ing”, IEEE network magazine, March 2000.
11. H. Mantar, I. Okumus, J. Hwang, S. Chapin ” An Intra-domain Resource Man-
agement Model for Diffserv Networks”, Journal of High Speed Networks, Vol. 15,
pp.185-2005, 2006.
An Admission Control and TXOP Duration of
VBR Traffics in IEEE 802.11e HCCA with
Guaranteed Delay and Loss
Tae Ok Kim1 , Yong Chang2 , Young-Tak Kim3 , and Bong Dae Choi1
1
Department of Mathematics and Telecommunication Mathematics Research Center,
Korea University, Korea
2
Telecommunication Systems Division, Information & Communication Business,
Samsung Electronics Co. Ltd, Korea
3
Dept. of Information and Communication Engineering, Graduate School,
Yeungnam University, Korea
[email protected], [email protected], [email protected],
[email protected]
1 Introduction
IEEE 802.11 wireless LANs(WLANs)[1] is one of the most popular wireless
technologies because of low cost and easy deployment. The explosive growth
of real time and multimedia applications arose the problem of the required qual-
ity of service(QoS) of these applications such as guaranteed packet delay and
packet loss probability. However 802.11 DCF protocol does not provide any QoS
support.
Hybrid Coordination Function(HCF) in IEEE 802.11e[2] is to enhance the
QoS Support by combining and extending the DCF and PCF of the MAC sub-
layer. The HCF consists of two channel access mechanisms : a contention-based
channel access(EDCA) providing a probabilistic QoS support and a controlled
channel access(HCCA) providing a parametric QoS support.
This paper is focused on a method to determine an optimal TXOP dura-
tion(TD) for the HCCA mechanism which provides a guaranteed loss probability
This research is supported by the MIC, under the ITRC support program supervised
by the IITA.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 162–169, 2006.
c Springer-Verlag Berlin Heidelberg 2006
An Admission Control and TXOP Duration of VBR Traffics 163
and delay. As we will describe in more details in the next section, the 802.11e
standard[2] proposes a reference design of the simple scheduler and admission
control that is efficient for traffic with Constant Bit Rate(CBR) characteristic.
However, a lot of applications such as video have VBR characteristic and the
reference design would produce large loss probability for VBR traffic[4].
Several improvements for taking account of guaranteed loss probability were
proposed. Ansel et al.[3] used a new scheduling algorithm called FHCF that
utilizes difference between estimated average queue length at the beginning of
the SI and ideal queue length. Fan et al. [4] proposed a scheme that decides
TXOP duration by inversion of packet loss probability formula. Even if this
scheme guarantees QoS, it has a tendency to overestimate TXOP duration.
In this paper, We propose a new method to determine the effective TXOP du-
ration (TD) of scheduling algorithm for VBR traffics that guarantees delay and
loss probability in IEEE 802.11e HCCA. In this algorithm, the TXOP Duration
for each Traffic Stream (TS) is determined by the number (N ) of packets to be
transmitted during a TXOP duration, which is found by analyzing M/DN /N × l
queueing system where the delay bound of the TS is l times the Service Inter-
val(SI) whose length is D. Numerical results show that the number of admitted
TSs with guaranteed loss probability by our new method is greater than the
number of admitted TSs by other known schemes.
The rest of the paper is organized as follows. Section 2 presents a reference
design of the simple scheduler and admission control for completeness. Section
3 presents our new algorithm for calculation of an effective TXOP duration,
Section 4 gives some numerical results to compare the other schedulers.
all Maximum Service Intervals for all admitted streams, which is a submultiple
of the beacon interval. This value will be the SI for all stations with admitted
streams.
In the second step, the TXOP duration for a given SI is calculated for the
stream. For the calculation of the TXOP duration for an admitted steam, the
scheduler uses the following parameters: Mean Data Rate(ρ), Nominal MSDU
Size(L) from the negotiated TSPEC, the Scheduled Service Interval(SI) calcu-
lated above, Physical Transmission Rate(R), Maximum allowable Size of MSDU,
i.e., 2304bytes(M ) and Overheads in time units(O). The Physical Transmission
Rate is the Minimum PHY Rate negotiated in the TSPEC. If Minimum PHY
Rate is not committed in ADDTS request, the scheduler can use an obseved
PHY rate as R. In order to indicate the parameters of a specific TS ”m”, para-
meters are put subindex ”m” like ρm and Lm . The TXOP duration is calculated
as follows. First, the scheduler calculates the number Nm of MSDUs arriving
during the SI for the traffic TS m with the Mean Data Rate ρm :
ρm ∗ SI
Nm = . (1)
Lm
Then the scheduler calculates the TXOP duration(T Dm) of TS ”m” as the
maximum of (1)time to transmit Nm frames at rate Rm and (2) time to transmit
one maximum size MSDU at rate Rm (plus overheads):
Nm · L m M
T Dm = max( +O , +O ) . (2)
Rm Rm
Fig. 2. Buffer state and behavior of packet loss in the reference design
calculates the number of MSDUs that arrive at the Mean Data Rate during
the Scheduled Service Interval. Second, the admission control unit calculates
the TXOP duration(TD) that needs to be allocated for the stream. Finally, the
admission control unit determines that the stream can be admitted when the
following in equality is satisfied:
T Dk+1 k
T Dm Tb − Tcp
+ ≤ (3)
SI m=1
SI Tb
where k is the number of existing streams and k + 1 is used as index for the
newly arriving stream. Tb indicates the superframe duration and Tcp is the time
used for EDCA during the superfrme.
Fig. 3. Buffer state and behavior of packet loss in our scheme when l = 2 (Nm =7)
Then we can choose Nm as the minimum number that the packet loss probability
is bounded by the required loss probability.
So, the TXOP duration for the TS is decided by
Nm · L m M
T Dm = max( +O , +O ) . (6)
Rm Rm
After decision of TXOP duration by this procedure, the admission control
process is done by the same way as the reference design.
4 Numerical Results
SIFS 10 us
PHY rate (R) 11 Mbps
Minimum PHY rate (Rmin ) 2 Mbps
From [5], the bit rate of the ordinary streaming video over Internet is from
300kbps to 1Mbps. In the numerical analysis, we consider three mean data rate of
video flow: 300kbps, 600kbps and 1Mbps. For each video stream of different mean
Reference Fan’s
scheme scheme[4]
ρ L(byte) N TD(ms) Pl N TD(ms) Pl
750 5 3.976 0.17547 10 7.953 0.00444
300 Kbps 1000 4 3.908 0.17228 8 7.817 0.00602
1250 3 3.477 0.2240 7 8.112 0.00573
750 10 7.953 0.12511 16 12.724 0.00547
600 Kbps 1000 8 7.817 0.11479 13 12.702 0.00529
1250 6 6.953 0.160623 11 12.748 0.00579
750 17 13.520 0.08789 23 18.291 0.00806
1 Mbps 1000 13 12.702 0.09397 18 17.588 0.00950
1250 10 11.589 0.12511 16 18.543 0.00547
168 T.O. Kim et al.
data rate, different packet size are used in the analysis. Nominal MSDU size of
each video stream are 750bytes, 1000bytes and 1250 bytes. The number of packet
arrived during SI of VBR video source is assumed to be Poisson distributed. The
video source is assumed to have a fixed packet length equal to Nominal MSDU
size. The SI is 100ms and the proportion of the contention-free period in a SI
T −T
is set to be half of the SI(i.e b Tb cp = 0.5). We assume that the packet loss
probability requirement is 1%
The numerical results for other schemes and our scheme is presented in table 2
and 3 (N = the number of packets to be transmitted during a TXOP duration),
An Admission Control and TXOP Duration of VBR Traffics 169
respectively. Table 2 quotes from [4]. It shows that the packet loss probability
of the reference scheme for the VBR traffic is quite high, and consequently the
reference scheme is not good for the VBR traffic. For proposed scheme in [4], the
packet loss probability is bounded by 1%. However the required TXOP duration
of this scheme is twice as long as one of the reference design. For our scheme,
the packet loss probability is bounded by 1% and the TXOP duration is remark-
ably shorter than that of proposed scheme in [4]. Note that Fan’s scheme[4] is
essentially the same as our scheme with l = 1.
Fig. 4 represents the the admission regions of the proposed scheme in [4] and
our scheme when there are two types of traffics, CBR traffics of 50ms delay bound
and VBR traffics of 100ms delay bound. By this figure, we can recognize the fact
that the admission region for our scheme is larger than one for the scheme in [4]
and is a little bit less than the reference design. Therefore our proposed scheme is
better than other two schemes in terms of packet loss probability and admission
region.
References
1. IEEE, ”International Standard for Information technology-Telecommunications and
information exchange between systems-Local and metropolitan area networks-
Specific requirements-Part 11:Wireless Medium Access Control(MAC) and Physical
Layer(PHY) specifications,” IEEE-802.11-1999, 1999
2. IEEE, ”IEEE Standard for Information technology-Telecommunications and in-
formation exchange between systems-Local and metropolitan area networks-
Specific requirements-Part 11:Wireless Medium Access Control(MAC) and Physical
Layer(PHY) specifications: Amendment 7: Medium Access Control(MAC) Quality
of Service(QoS) Enhancements,” IEEE P802.11e/D11.0,Oct. 2004
3. P. Ansel, Q. Ni and T. Turletti, ”FHCF: A Fair Scheduling Scheme for 802.11e
WLAN,” INRIA Research Report No 4883, Jul. 2003
4. W.F. Fan, D.Y. Gao, D. H.K. Tsang and B. Bensaou, ”Admission Control for Vari-
able Bit Rate traffic in IEEE 802.11e WLANs,” to be appeared in The Joint Con-
ference of 10th Asia-Pacific Conference on Communications and 5th International
Symposium on Multi-Dimensional Mobile Communications, Aug. 2004
5. F. Kozamemik, ”Media Streaming over the Internet-an overview of delivery tech-
nologies,” EBU Technical Review, Oct. 2002
NETSAQ: Network State Adaptive QoS Provisioning
for MANETs
1 Introduction
The role of Mobile Ad-hoc networks (MANETs) [1], in realization of a ubiquitous
world, is manifold. MANET devices can serve as user interface, provide a distributed
data storage, or act as a mobile infrastructure access point for other nodes. The
MANETs are infrastructure-less networks that are envisioned to be spontaneously
created whenever two or more nodes come in close proximity to each other. MANETs
are characterized by dynamic topologies, limited processing and storage capacity, and
bandwidth constrained wavering capacity links.
The described inherent characteristics of MANETs implicate newer requirements
and technical challenges for the management of such networks. The incorporation of
modern real time services, like transfer of audio and video concomitant to delay-
agnostic services has further increased the management complexity. On one hand,
MANET nodes have limited resources to share amongst contending services, on the
other hand such services need and expect high priority on the network in order to
meet higher user expectations about their reliability and quality.
*
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 170 – 179, 2006.
© Springer-Verlag Berlin Heidelberg 2006
NETSAQ: Network State Adaptive QoS Provisioning for MANETs 171
2 Related Work
Ensuring QoS through routing is relatively a new issue when it comes to MANETs.
QoS routing protocols aim to search for routes with sufficient resources to satisfy
initiating applications. The QoS routing protocols work closely with the resource
management module in order to set up paths through a list of mobile nodes to meet
the end-to-end service requirements in terms of bandwidth, delay, jitter, and loss, etc.
The computational and communication cost of QoS routing is known to be fairly high
and it has raised the questions whether or not should it be tackled in MANETs.
Many proposals have been made for QoS provisioning in MANETs. The In band
Signaling (INSIGNIA) [3], Stateless Wireless Ad hoc Networks (SWAN) [4], Core
Extraction Distributed Ad Hoc Routing (CEDAR) [5], Adaptive Source Routing
(QoS-ASR) [6], and Quality of Service for Ad hoc Optimized Link State Routing
Protocol (QOLSR) [7] are examples of QoS-routing proposals. We have summarized
172 S.A. Chaudhry et al.
various parameters to have an insight into the important features provided by these
protocols. Table 1 describes the distinguishing features of these protocols.
Notify Policies
Event Event Repository
Generation Policy Parser
Register OS
Event
Policy Manager
Monitoring Condition Evaluater
Monitoring and Reporting Action
Monitoring and
Reporting Agent
Policy Enforcer
(b) ELA for normal MANET node
Configuration Agent Config.
Filtering Aggregation Action
Module Module Local Mgmt. Services
Legend
Policy Distributor
Config.
Settings MCEE : Mobile code execution environment
Local Local
Mgmt. Low level MIB : Management information base
Monitoring services to policies to
Logs Config. nodes ELAs
SLA : Simple local agent
Local Configuration and Management Database ELA : Extended local agent
DPM : Domain policy manager
Polling SLAs Context from SLAs and ELAs
Fig. 1. Management components for cluster head and normal MANET node
Consider a network that provides services to users that belong to different groups such
as special and ordinary groups. Now a MANET special user starts a video
conferencing application. The QoS of the application can be violated due to: a)
mobility, whenever the source, the destination or any intermediate node moves, the
established QoS path can be broken, b) interference or contention, whenever the
number of nodes around the QoS path increase considerably, interference increases
and contention against wireless channels escalates. In both the situations there is a
possibility of QoS degradation even if the resources were reserved beforehand. In
either case, the alternative path is needed which can facilitate the required QoS
provisioning. This phenomenon is inherent to MANETs, and cannot be eliminated.
requirements with minimum and maximum QoS bandwidth constraints bmin kbps and
bmax kbps. In order to admit this application into the network, the average data rate of
the network is checked. A bandwidth threshold s, based on network media access
delay D and Packet Loss Ratio (PLR) is calculated for the application at the initiating
node of the application. The resources are then marked along the path based on s, i,
and if such bandwidth is not available then we use bmin. Once all the links along the
path, from source to destination are covered, then we start the application flow. After
the flow is established, resources can be reserved on intermediate nodes according to
threshold s. This, posteriori reservation, is different from reserving the resources
before starting the flow (priori reservation). Posteriori reservation eliminates the
complexity and delay which is inherent to priori reservation and provides better
response time to the application user. The application flow is re-routed if the end-to-
end bandwidth falls below the reserved resources for more than a specific time. The
algorithm is depicted in fig. 4.
If the established path is broken due to link failure or mobility of any node(s),
rerouting is done at the node where the path is broken. If this intermediate node can’t
find the QoS path, the next upstream node tries to find the path. If no QoS path can be
established from the upstream node as well, source is notified to broadcast a new
route request with bmin kbps and bmax kbps QoS constraints.
If the data rate for the application decreases than the reserved bandwidth for more
than t sec, the destination will notify the source. It means all the intermediate nodes
do not need to continuously monitor the data rate violations. Also, in case of
interference, the source has a greater possibility to find new disjoint routes. The
routing agent at the source will then be invoked to find a new route with data rate > s
kbps or > i kbps or > bmin kbps as it would be described in the policy.
MANET routing protocols may use link level acknowledgement (ACK) messages
and a timeout period for link level connectivity information. In an area with dense
NETSAQ: Network State Adaptive QoS Provisioning for MANETs 175
population of nodes, hidden terminal problem can become quite significant. Due to
hidden terminal problem and high contention, some nodes will not receive the link
layer ACK packets from the neighboring nodes. When the timeout period expires, a
node declares the link as broken, discards all ongoing communication packets and
generates a route error message [10]. This causes the throughput to drop drastically.
The communication resumes when a new path is found or the same path is eventually
re-discovered. This instability problem is caused by fast declaration of link failures
which is rooted at the link layer. The breaking and rediscovery of the path result in the
drastic throughput oscillations. In order to avoid this problem, we extend the solution
proposed in [11]. This solution uses a “don’t-break before-you-can-make” strategy.
This strategy is based on modifying the routing algorithm so that the routing agent
continues to use the previous route for transmissions before a new route can be found.
When the new route is found or the same route is eventually re-discovered, all nodes
discard the previous route and switch to the new one (or the same one) for
transmissions. An example for the explanation is shown is fig.3.
Let bmin = 50 kbps, bmax = 100 kbps, t = 1 min, w1D = 0.2, w2PLR = 0.3 and i=65 kbps
Then threshold s:
s = max [{1- 0.5 (w1 D + w2 PLR)} bmax, bmin]
s = max [{1- 0.5 (0.2 + 0.3)} 100, 50] = 75 kbps
Reserve the resources along the path equal to 75 kbps (s)
Else reserve the resources equal to 65 Kbps (i)
Otherwise reserve the resources equal to 50 Kbps (bmin)
If resources are reserved equal to 75 kbps
when destination node receives data rate < 75 kbps for 1 min
It will notify the source
Source will establish a new route with data rate > s kbps
If no such route is available, a route with data rate > i or minimum bandwidth bmin condition will be set up
If resources are reserved equal to 65 kbps
when destination node receives data rate < 65 kbps for 1 min
It will notify the source
Source will establish a new route with data rate > s kbps
If no such route is available, a route with data rate > i or minimum bandwidth bmin condition will be set up
If resources are reserved equal to 50 kbps
when destination node receives data rate < 50 kbps for 1 min
It will notify the source
Source will establish a new route with data rate > s kbps
If no such route is available, a route with data rate > i or minimum bandwidth bmin condition will be set up
5 Performance Evaluation
The simulations is this section evaluate the suitability of the algorithm to support
adaptive flows in a MANET under various conditions such as traffic, mobility, and
channel characteristics. In particular, we evaluated system wide adaptation dynamics
and the impact of threshold based rerouting mechanisms and mobility on end-to-end
sessions.
176 S.A. Chaudhry et al.
We measure the performance of two adaptive flows with User Datagram Protocol
(UDP) and Transmission Control Protocol (TCP). Fig. 4 and 5 show flows with bmax
values as 160 and 320 while bmin values are 90 and 160 respectively.
160
140
Throughput (Kbps)
120
100
UDP Flow 1
80
TCP Flow 1
60
40
20
0
0
108
231
342
462
576
693
810
924
1044
1155
1278
1386
1494
1617
1728
1890
Time (sec)
350
300
Throughput (Kbps)
UDP Flow 2
250
TCP Flow 2
200
150
100
50
0
0
105
210
315
420
525
630
735
840
945
1050
1155
1260
1365
1470
1575
1680
1785
1974
Time (sec)
Variable traffic load was generated at random nodes that caused the flows to be
rerouted depending on the dynamic network conditions. Momentary QoS violations
occurred due to rerouting but overall the flow kept its minimum bandwidth guarantee.
We observed that the flow fluctuated between its maximum and minimum bandwidth
constraints. Especially the flows under TCP fall below minimum bandwidth
constraint at times. This is due to the TCP rate adjustment because of large network
delay and packet loss.
250
200 200
Throughput (Kbps)
Throughput (Kbps)
Throughput (Kbps)
200 Normal_stream
150 150
Normal_stream1 1
video_streaming1
150
100Normal_stream1 Normal_stream video_streaming2
100
100 2
50
50 50
0
0 0
0 100 200 300 400
0 100 200 300 400 -50 0 100 200 300 400
-50
-50
Time (min) Time (min)
Time (min)
(a) 5 nodes (b) 10 nodes (c) 15
(c)nodes
15 nodes
T hro u gh p u t (K b p s)
T hro u gh p u t (K b p s)
50 50 50
0 0 0
0 100 200 300 400 0 100 200 300 400 0 100 200 300 400
-50 -50 -50
Time (sec) Time (min) Time (min)
(a) 5 nodes (b) 10 nodes (c) 15 nodes
The Fig. 7 shows NETSAQ adapts to the changing conditions and finds a route with
a slightly decreased QoS value within the minimum and maximum QoS constraints for
both the streams. NETSAQ reroutes both the streams to a lower QoS levels to cope up
with the degrading network conditions, but when the loading traffic increases and
saturates the network towards the end, only best effort QoS can be provided.
We evaluated the performance under increasing transmission rates (by changing the
air interface and increasing stream rates as well. The simulation environment consists
of 15 ad hoc nodes in an area of 200m x 200m. Each mobile node has a transmission
range of 100m. The transmission rate or air interface is changed from 1 to 2 and 5.5
Mbps. The comparison between Fig. 8 and Fig. 9 allows us to observe the way
NETSAQ adapts to transmission rate variations.
600
300
Normal_stream1 Normal_stream1
150 250
Normal_stream1 Normal_stream2
400 Normal_stream2
200
Normal_stream2
100 Normal_stream3 Normal_stream3
150
200
50 100
50 0
0
0 0 100 200 300 400
0 100 200 300 400
-50 -50 0 100 200 300 400 -200
T h ro u g h p u t (K b p s )
600
250 NETSAQ_stream1 NETSAQ_stream1
150 500
NETSAQ_stream1
200 NETSAQ_stream2 NETSAQ_stream2
NETSAQ_stream2 400
100 150 NETSAQ_stream3 NETSAQ_stream3
300
100
50 200
50 100
0 0 0
0 100 200 300 400
-50 -50 0 100 200 300 400 -100 0 100 200 300 400
Time (min) Time (min) Time (min)
(a) 1 Mbps (b) 2 Mbps (c) 5.5 Mbps
6 Conclusion
In this paper we have proposed NETSAQ, a network state adaptive QoS algorithm for
MANETs. NETSAQ implements the QoS provisioning policies using any underlying
NETSAQ: Network State Adaptive QoS Provisioning for MANETs 179
routing protocol. It eliminates the bulky computation and communication for QoS in
routing algorithm. It utilizes lesser MANET resources as compared to existing QoS
routing protocols. It is simple yet minimizes the degradation of best effort traffic,
which is observed in many QoS routing schemes.
References
1. Frodigh, M., Johansson, P., Larsson, P.: Wireless ad hoc networking: the art of networking
without a network. Ericsson Review, No.4, pp. 248-263, 2000.
2. Veres, A., Campbell, A., Barry, A., Sun, L. H.: Supporting Service Differentiation in
Wireless Packet Networks Using Distributed Control. IEEE JSAC, vol. 19, no. 10, Oct.
2001.
3. Lee, S., Ahn, G., Zhang, X., Campbell, A.: INSIGNIA: IETF Internet Draft, draft-
ietfmanet-insignia-01.txt, work in progress, Nov. 1999.
4. Ahn, G., Campbell, A., Veres, A., Sun, V.: SWAN, IETF Internet Draft, draft-ahn-
swanmanet-00.txt, work in progress, Oct. 2002.
5. Sivakumar, R., Sinha, P., Bharghavan, V.: Core Extraction Distributed Ad Hoc Routing
(CEDAR) Specification. IETF Internet Draft, draftietf-manet-cedar-spec-00.txt, work in
progress, 1998.
6. Labiod H., Quidelleur, A.: QoS-ASR: An Adaptive Source Routing Protocol with QoS
Support in Multihop Mobile Wireless Networks. IEEE VTC’02, pp. 1978- 1982. 2002.
7. Badis, H., Al-Agha, K.: Quality of Service for Ad hoc Optimized Link State Routing
Protocol (QOLSR). IETF Internet Draft, work in progress draft-badis-manet qolsr-01.txt,
Apr. 2005
8. Ge, Y., Kunz, T., Lamont, L.: Proactive QoS Routing in Ad Hoc Networks. ADHOC-
NOW’03, pp. 60-71, 2003.
9. Chaudhry, S.A., Akbar, A.H., Siddiqui F.A., Sik Y.W.: Autonomic Network Management
for u-Zone Networks. UbiCNS’ 05, Jeju, Korea, June 9-10, 2005
10. Charles E. Perkins and Elizabeth M. Royer, “The Ad hoc On-Demand Distance Vector
Protocol” In Charles E. Perkins, editor, Ad hoc Networking, pages 173–219. Addison-
Wesley, 2000.
11. Chung-Ng, P. Liew, K.: Re-routing Instability in IEEE 802.11 Multi-hop Ad-hoc
Networks. WLN'04, Nov. 2004.
End-to-End QoS Guaranteed Service in WLAN
and 3GPP Interworking Network
1 Introduction
In recent years, many mobile users are demanding anytime and anywhere access
to high-speed multimedia services for next generation communication system.
There are many number of communication technologies for next generation sys-
tem. In order to satisfy the user requirements for the wireless local area network
(WLAN) and third generation partnership project (3GPP) interworking net-
work, the 3GPP is concerned about the WLAN and 3GPP interworking [1].
3GPP has been studying and standardizing the WLAN and 3GPP interwork-
ing mechanism. However, it is insufficient to investigate the quality of service
(QoS) provisioning technology in the WLAN and 3GPP interworking network[2].
There are various challenges to provide the end-to-end QoS guaranteed ser-
vices through WLAN and 3GPP interworking network [3], [4]. First, there are
many differences between their QoS provisioning technologies such as the QoS
parameters, service classes and so on. Accordingly, the mapping mechanism is
required for a seamless service. Second, a bottleneck may be generated due to
the limited capacity and the overload at a gateway linked with backbone net-
work. In order to solve these problems, we define the functional features based
on the end-to-end QoS architecture in WLAN and 3GPP interworking network.
We also propose the new QoS provisioning technologies, and then analyze the
performance of the proposed mechanism using a simulator.
1
This work was supported by the second stage of Brain Korea 21 (BK21) Project in
2006.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 180–189, 2006.
c Springer-Verlag Berlin Heidelberg 2006
End-to-End QoS Guaranteed Service in WLAN and 3GPP 181
In the conventional WLAN and 3GPP interworking standard, the QoS provi-
sioning mechanisms have not been considered in detail [2]. Therefore, we define
the functional features for the end-to-end QoS provisioning based on the end-
to-end QoS architecture and model the control and user plane considering the
detailed functional features. Besides the control and user plane, we propose QoS
parameter/class mapping and packet scheduler (DPS).
differentiated service (DiffServ) and integrated service (IntServ). When the Diff-
Serv mechanism is applied to the backbone network, the PDG has to perform a
DiffServ edge function.
Based on the end-to-end QoS architecture, we model control and user plane as
shown in Fig. 3 and Fig. 4. In control plane, we define the functional blocks
which manage the signal procedure for the resource reservation. In user plane,
we define the functional blocks that are needed for QoS guaranteed service when
packets are transmitted through the WLAN and 3GPP interworking network.
The definition of the functional features is as follows.
For the QoS provisioning, we propose the QoS parameter/class mapping tech-
nology and DPS packet scheduler based on the control plane and user plane.
The QoS parameter and class mapping technologies are used to translate the
184 S.-M. Oh et al.
QoS information in the control plane and user plane. The packet scheduler is
applied to manage the data flow in user plane. We evaluate the performance of
the QoS parameter/class mapping algorithm and packet scheduler in section 4.
QoS Service Class Mapping. We should consider QoS service class mapping
among WLAN, IEEE 802.1D (IP) and 3GPP for the packet scheduling, because
the data packet would be transmitted through WLAN, IP, and 3GPP network.
End-to-End QoS Guaranteed Service in WLAN and 3GPP 185
Class 1
CAC
Class 2 Reservation
Class 4
Weighted Round
Robin
Fig. 5. DPS
Since the QoS service classes of them are defined for each service requirement,
we can make the QoS service class mapping table considering the similar service
requirements like Table 2.
Dynamic Processor Sharing (DPS). The DPS is proposed for the QoS
guaranteed service. DPS can keep the delay bound of voice traffic by the resource
allocation for voice traffic. In addition, DPS can provide the fairness for the
others. The DPS is shown in Fig 5 and the explanation of DPS operation is as
follows. We define the four service classes as class1, class2, class3, and class4
according to the priority. Class1 indicates the highest priority traffic class which
is very sensitive to delay such as voice of internet protocol (VoIP). To support
the class1, the resource manager allocates the resource to guarantee the delay
bound when the class1 is generated. Class2 is also sensitive to delay, but the
tolerance of delay is larger than class1 like video streaming. Class3 and class4
are insensitive to delay, but they are critical to the packet drop probability
such as hyper text transfer protocol (HTTP) and file transfer protocol (FTP).
Therefore, the weighted round robin (WRR) is applied to the class2, class3,
186 S.-M. Oh et al.
class4 to keep the QoS and fairness. The weight should be selected considering
the traffic load for the efficiency of capacity. Since class1 occupies the resource,
the fairness problem may be occurred if the resource limitation is not defined.
To solve this fairness problem, the CAC is applied to the class1.
4 Performance Evaluation
In this section, we present the network model and results of the performance
evaluation for our proposed QoS provisioning technologies. We built the simula-
tor for the QoS provisioning technologies using OPNET.
0.1
voice delay of DPS
voice delay of FIFO
0.06
0.05
0.04
0.03
0.02
0.01
0
1 2 3 4 5
Number of voice users
Fig. 6. End-to-end delay of voice and video traffic according to the packet schedulers
and the number of voice users
0.14
voice delay of DPS
voice delay of FIFO
0.08
0.06
0.04
0.02
0
1 1.3 1.5 1.7
PHY throughput of video traffic (Mbps)
Fig. 7. End-to-end delay of voice and video traffic according to the packet schedulers
and the video traffic load
0.0665
video delay of SP
video delay of DPS
0.066
video end−to−end delay(sec)
0.0655
0.065
0.0645
0.064
0.0635
0.063
0.0625
0.062
1 2 3 4 5
Number of voice users
Fig. 8. End-to-end delay of video traffic according to the number of voice users
end-to-end delay of video steaming, FTP and HTTP traffic may increase because
of the unfairness of SP. Unlike SP, DPS is able to guarantee the fairness because
of the CAC applied to DPS. For example, if we limit the maximum number of
voice user in DPS, DPS can drop the voice call when the number of voice users
is over the maximum number. Fig. 8 presents the video end-to-end delay when
the maximum number of voice users is equal to 3. DPS guarantees the QoS for
a video streaming even though the number of voice users rapidly increases.
End-to-End QoS Guaranteed Service in WLAN and 3GPP 189
5 Conclusion
In this paper, we defined the functional features of the network elements and
designed the control plane and user plane considering the end-to-end QoS ar-
chitecture. We proposed the QoS parameter/class mapping and DPS to provide
the end-to-end QoS guaranteed service. We also built the simulation model and
verified the performance of the proposed technologies. The simulation results
show that DPS is superior to FIFO, WRR and SP for the multimedia services
such as voice and video.
Since the WLAN and 3GPP interworking technology is the newly embossed
issue, our proposed end-to-end QoS mechanisms are expected to be a significant
reference. Contributions of this paper are as follows.
– Design the control and user plane
– Propose the QoS parameter/class mapping and DPS
– Build the simulation model using OPNET for the WLAN and 3GPP inter-
working network
Finally, this framework can be applied to the other heterogeneous network be-
tween WiMAX and 3GPP.
References
1. W. Zhuang, Y. S. Gan, K. J. Loh, and K. C. Chua, “Policy-based QoS management
architecture in an integrated UMTS and WLAN environment,” IEEE Commun.
Mag., vol. 41, pp. 118-125, Nov. 2003
2. “QoS and policy aspects of 3GPP - WLAN interworking (Release 7),” 3GPP TR
23.836 v0.4.0, 2005
3. D. Skyrianoglou and N. Passas, “A framework for unified IP QoS support over
UMTS and WLANs,” in Proc. European Wireless 2004, Barcelona, Spain, Feb. 2004
4. S. I. Maniatis, E. G. Nikolouzou, and I. S. Venieris, “QoS issues in the converged
3G wireless and wired networks,” IEEE Commun. Mag., vol. 40, pp. 44-53, Aug.
2002
5. “3GPP system architecture evolution : report on technical options and conclusions,”
3GPP TR 23.882 v.0.1.1, 2005
6. “3GPP system to WLAN interworking; system description,” 3GPP TS 23.234
v6.4.0, 2005
7. “QoS concept and architecture (release 4),” 3GPP TS 23.107 v4.4.0, 2002
8. “Amendment : MAC QoS enhancements,” IEEE P802.11e/D13.0, Jan. 2005
Network-Adaptive QoS Routing Using Local
Information
Jeongsoo Han
1 Introduction
In localized routing technique, each node in the network makes local routing
decisions based just on local information, i.e., information on the status of the
network that has been obtained from the neighbor nodes. Also, routing policies
aim to minimize the average overall blocking probability based on local informa-
tion. If local information is used, it is difficult to determine how a routing decision
made on a single node may influence the network’s overall performance. It has
been demonstrated that localized QoS routing is simple, stable, adaptive and
effective in comparison to global QoS routing schemes[1]. [2,3] have proposed a
localized QoS routing scheme, proportional sticky routing(psr). The psr scheme
maintains self-adaptivity by using the maximum permissible flow blocking para-
meter and controlling the number of flows routed along a path in each cycle and
by re-adjusting flow proportions after every observation period. But, this ap-
proach has three main drawbacks. Firstly, the traffic pattern is always unknown
in practical networks. Secondly, we need to know the precise information along
the path to calculate the blocking probability. Thirdly, even though we are able
to identify the traffic pattern and calculate the blocking probability, the iteration
time needed to solve the global optimization problem can be significant[4]. To
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 190–199, 2006.
c Springer-Verlag Berlin Heidelberg 2006
Network-Adaptive QoS Routing Using Local Information 191
2 Proposed Algorithm
In localized routing technique, each node in the network makes local routing de-
cisions based just on local information. From this viewpoint, it can be connected
with POMDP problem which its agent ignores about its environments. That is,
that the agent knows a) its own state xt (n) ∈ X at all times n within trial t, b)
the costs Cx (a) of taking the actions, c) the states that are absorbing, and d)
the distribution governing the way the transition probabilities change. But, the
agent does not know transition probabilities and its knowledge as to how they
change. This ignorance turns a Markov decision problem into a POMDP.
192 J. Han
i
t ,n t
q (r1 )
AB E[ p (r1 )]
AB
r1
C A (r1 ) 2
q tAC
,n
(r3 ) E[ p tAC (r3 )]
Under the approximation, the agent performs value iteration in the assumed
mean process as follows:
Network-Adaptive QoS Routing Using Local Information 193
t,n
Vα+1 (x, y) = min{Cxy (a) + γΣy∈X qxy
t,n
(a)Vαt,n (x, y)} (1)
a∈A
The course of updating routing information is as follows : Once the agent picks
one of the possible actions that minimizes the right-hand side of equation (1),
performs it, uses Bayes’ rule to compute the posterior distribution over the
transition probabilities based on the transition actually observed, and calculates
t,n+1
qxy (a). The agent attempts to take such an action for admitting a request
with certain bandwidth requirement if there is enough bandwidth in the chosen
action, so it is possible to route a request on that action. In this paper, we define
that such an action is termed as effective action and is defined as et,n xy (a) = 1.
And if otherwise, ineffective action and et,n xy (a) = 0. In this paper, the agent
does not know et,nxy (a), ie it does not know whether it is possible to route on such
an action because it does not absent of global information. Therefore, we define
a probabilistic model(φ) of the efficacy of the transitions ie, the probabilities of
t,n
routing success on each path. Specifically, let qxy (a)be the agent’s estimate of
the efficacy of action a at x for y at time n during trial t. And the agent assumes
that between each trial with some small arrival rate(κ), each et,n xy (a) gets set to
a new probability value, independent of its previous value. After trial t − 1, the
updating routing information at x occurs as follows:
⎧
⎪
⎨κφ + (1 − κ)qx
t−1,0
(a) (2-1),
qx (a) = 1 − κ(1 − φ)
t,0
(2-2), (2)
⎪
⎩
κφ (2-3).
When a was not tried at x during the trial, we can define as (2-1). When a
was tried at x and was successful, we also define as (2-2). And (2-3) means when
t,n
a was tried at x and was unsuccessful. qxy (a) is reset when the agent tries a at
t,n
x to whatever actually happened. For actions that were not attempted, qxy (a)
relaxes towards φ at a rate governed by the arrival rate κ. And others cases,
t,n
qxy (a) is reset as each values.
The decision-making process consists of selecting the path through which the
request will be connected. The agent’s initial uncertainty about the transition
or reward structure of the world should drive the initial experimentation, and, if
the world can change stochastically over time, then this further source of uncer-
tainty should drive continuing exploration. Unfortunately, it is computationally
intractable for the agent to determine the correct experiments even if it knows
what it does not know. And in using this mean process, the approximation fails
to directly account for the fact that if a transition on the way to the destination
194 J. Han
is blocked because of QoS problem, then the agent will have to choose some
other path to that destination. To solve this problem in this paper, we shows the
heuristic algorithm, exploration bonus method which is added to the immediate
value of a number that is a function of this how long it has been since the agent
has tried that action in that state. Therefore, rather than have the agent model
the efficacy of transitions, we had
it directly add to the immediate cost for a
move an exploration bonus of α nx (a) for trying action a at state x, if that
action had not been tried for nx (a) trials.
1 4 7 10 12
2 5 8 13
3 6 9 11 14
(a) Simple Test Network
13 15
5
9 17
2
4 7 18
10
12 16
8
1 3 14
11
6
(b) The ISP Network
10
9
8
average hop counts
7
6
5
4
3
2
1
0
0.1 0.3 0.5 0.7 0.9 1
Fig. 3. The performance of routing average hop count on test network which varies
with φ value
Fig. 3 shows the performance of routing average hop count on each test net-
work which varies with a φ value. As φ gets bigger, the system will always
eventually explore to find paths that might be longer than its current one - ie it
will find the detours when the current path will be blocked because of various
QoS parameters. But as φ gets smaller, the system becomes more pessimistic.
It will not bother to find large detours to try possibly better paths. Therefore,
the lower φ, the more restricted the excursions. As seen in Fig. 3, the actual
average length of the shortest paths also increased(from 3.5 hop counts at the
real value of φ = 0.2 to 5 hop counts at the real value of φ = 0.9 in the Fig. 2
(a) test network). It is attributed to the fact that it will find detours other than
the current path as φ increases(κ was fixed at 0.03 and was known to the agent).
Also the Fig. 2 (b) test network that has relatively many links between the nodes
has more hop counts than the Fig. 2 (a) test network.
Network-Adaptive QoS Routing Using Local Information 197
10
9
8
average hop counts
7
6
5
4
3
2
1
0
0.1 0.3 0.5 0.7 0.9 1
Fig. 4. The performance of routing success rate on test network which varies with φ
value
60
50
Service Success rate(%)
40
30
20
10
0
0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05
k value
(a) Network (b) Network
Fig. 5. The performance of routing success rate on test network which varies with κ
value
Fig. 4 shows the performance of routing success rate on each test network
which varies with a φ value. As seen in the result of Fig. 3, the more φ increases,
the more detours other than the current path the agent explore. Therefore, the
performance of routing success rate will be improved. Also κ was fixed at 0.03
and was known to the agent.
Fig. 5 shows the performance of routing success rate on each test network
which varies with a κ value. As indicated above, the other parameter κ also has
an effect on the relative amounts of exploration and exploitation, but in a more
indirect way. Whereas φ determines the ultimate amount of exploration about
198 J. Han
the current best-known path, κ controls how soon that exploration happens.
Therefore, the more frequent explorations, the higher service success rate because
it will be clear that increasing κ increased the amount of exploration. φ was fixed
at 0.5. As seen in Fig. 5, the actual average success rate also increased (from
38% at the real value of κ = 0.03 to about 50% at the real value of κ = 0.05
in the Fig. 2 (b) test network). It is attributed to the fact that the amount of
exploration will increase.
4 Conclusion
In this paper, we present the new Localized Adaptive QoS routing technology
without the knowledge of network global information. For this, we proposed the
POMDP model and CEA technology that provide a better QoS routing method
using only local information on source nodes and provide a much more adaptive
way of path selection depending on a network environment. We also proposed an
edge-disjoint multi-path searching algorithm, SEMA. As with POMDPs, we start
from a probabilistic characterization of the uncertainty the agent has about the
network environment, apply this model to the Localized QoS routing method.
And we make a use of CEA, which use the mean values of random variables,
because performing dynamic programming to solve a POMDP is highly com-
putationally expensive and is intractable to determine the optimal solution. For
better path selection, we propose Exploration Bonus because network environ-
ment connected with agent can change over time, so a current optimal action
is not guaranteed at future. Also we evaluate performances of service success
rate and average hop count which varies with φ and κ performance parameters,
which are defined as exploration count and intervals. As a result, whereas the
parameter φ determines the amounts of exploration to find a better path than
the current one, κ controls how soon that exploration happens. As φ gets bigger,
the system will always find the detours when the current path will be blocked
because of various QoS parameters. But as φ gets smaller, the system becomes
more pessimistic. It will not bother to find large detours to try possibly better
paths. Therefore, if φ gets closer to 1, the success rate and the average hop
counts will get higher. Also, the higher κ, the higher service success rate because
the amounts of exploration will increase.
References
1. X.Yuan and A.Saifee, ”Path Selection Methods for Localized Quality of Service
Routing”, Technical Report, TR-010801, Dept of Computer Science, Florida State
University, July, 2001
2. Srihari Nelakuditi, Zhi-Li Zhang and Rose P.Tsang, ”Adaptive Proportional Rout-
ing: A Localized QoS Routing Approach”, In IEEE Infocom, April 2000.
3. Srihari Nelakuditi, Zhi-Li Zhang, ”A Localized Adaptive Proportioning Approach
to QoS Routing”, IEEE Communications Magazine, June 2002
Network-Adaptive QoS Routing Using Local Information 199
4. Y.Liu, C.K. Tham and TCK. Hui, ”MAPS: A Localized and Distributed Adaptive
Path Selection in MPLS Networks”in Proceedings of 2003 IEEE Workshop on High
Performance Switching and Routing, Torino, Italy, June 2003, pp.24-28
5. Yvn Tpac Valdivia, Marley M.Vellasco, Marco A. Pacheco ”An Adaptive Net-
work Routing Strategy with Temporal Differences”, Inteligencia Artificial, Revista
Lberoamericana de Inteligencia Aritificial, No 12(2001), pp. 85-91
6. Leslie Pack Kaelbling, Michael L. Littman, Andrew W.Moore, ”Reinforcement
Learning:A Survey”, Journal of Artificial Intelligence Research 4, 1996, pp 237-
285
7. P.Marbach, O.Mihatsch, and J.N.Tsitsiklis, ”Call Admission Control and Rout-
ing in Integrated Service Networks Using Neuro-Dynamic Programming”, IEEE
Journal on Selected Areas in Communications, Vol. 18, No.2, Feb 2000, pp.197-208
8. Jeong-Soo Han, ”A Localized Adaptive QoS Routing using TD(λ) method”, Jour-
nal of Korean Institute of Communication and Sciences, Vol.30, No.5B, pp. 304-309,
2005
9. Gregory Z. Grudic, Vijay Kumar, ”Using Policy Gradient Reinforcement Learning
on Automous Robot Controllers”, IROS03, Las Vagas, US, October, 2003 [11]
Richard S. Sutton etc, ”Policy Gradient Methods for Reinforcement Learning with
Function Approximation”, Advances in Neural Information Processing System, pp.
1057 1063, MIT Press 2000
10. S.Banerjee, R.K. Ghosh and A.P.K Reddy, ”Parallel algorithm for shortest pairs
of edge-disjoint paths”, J.Parallel Distrib. Comput. 33(2):165-171 (1996)
Configuration Management Policy in QoS-Constrained
Grid Networks
Institute (ETRI)
161 Gajeong-dong, Yuseong-gu, Daejeon, 305-700, Korea
[email protected]
1 Introduction
The global growth of Grid computing means the growth in application part reflecting
the needs of users. The traditional Grid computing (Computational, Data, Access) [1]-
[3] is more specialized, and consist of very various and complex grid services. All the
more Resource management need to correspond with the need of users or application
and policy based resource management system is one of the most promising candi-
dates for solving requirement of resource management.
For the reliable management of physically distributed and heterogeneous resources,
policy-based management [4] has been suggested. The connection between lower
network and the resource management system is mandatory for such a grid policy-
based resource management system. In other words, it is requisite to make the re-
source management system abstract because it is hard to manage heterogeneous lower
management system managed by local policy of different subject like a grid plane.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 200 – 209, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Configuration Management Policy in QoS-Constrained Grid Networks 201
However, this [4] isn’t consider about network resource management of lower net-
work layer such as L3/L2/L1. Actually, since the high performance and throughput
and QoS-constraint service requested with a specific Service Level Agreement (SLA)
can be supported throughout Grid service networks, the network resource manage-
ment is also considered importantly. For this network resource management, there are
many studies [8]-[10]. Especially, for high throughput issue, optical network is
mainly considered. [10] Moreover, as mentioned in [11], optical network with using
GMPLS give more efficient management and control scheme for Grid and network
resource management in Grid service networks.
In this paper, we propose the integrated grid resource management system architec-
ture based on Quality of Service (QoS) constraint configuration management policy in
Grid networks. This QoS-constraint configuration management policy is based on the
cost sensitivity using the cost according to providing network resources. Thus, we
define the cost function and cost sensitivity function for guaranteeing the QoS in
using network resources in this paper, and then, we derive the configuration manage-
ment policy rule. Also, we propose the policy enforcement process using this policy
rule in proposed grid resource management system architecture which is possible to
implement in Grid networks. Finally, through the theoretical and experimental analy-
sis, we can show that the proposed scheme outperforms the conventional scheme in
network performance and cost performance.
2 Related Works
The Policy Quorum based Resource Management Architecture is a kind of grid re-
source broker and scheduler that manage resources and jobs in the virtual Grid envi-
ronment based on Globus Core.[4] Policy quorum, be generated finally, represents the
collection of the resources that is adapted by individual policies according to private
request. Thus a user is satisfied with QoS by Policy Quorum. And several papers have
showed the resource reconfiguration algorithm based on temporal execution time
estimation method. Resource reconfiguration performs the reshuffling of the current
available resource set for maintaining the quality level of the resources. However
there is no consideration of the cost when reallocation job.
The policy-based Grid management middleware sits on the top of Gird Supporting
Environment and serves as the first access point for Gird administrator, or software on
its behalf, to configure and manage intra-domain and inter-domain Grid resources.
In order to deploy Policy-based management technology in Grid architecture, a
policy-based management standard architecture of the Internet Engineering Task
Force (IETF) can be used. These policy based management (PBM) system is origi-
nated from network management groups to reduce the complexity of management in
terms of QoS guarantees. It is suitable to complex management environment such as
large scale heterogeneous management in Grid networks. Quorum is a collection of
the elements which always intersects. It has been used to maintain the consistency of
replicated data, mutual exclusion. In PQRM the intersection property of Quorum is
used to make complex resource elements abstracted Grid resource which satisfies
various QoS requirements. The main result of PQRM is that QoS-constraint policy
202 H. Song et al.
requested. In this section, we define operation cost function and cost sensitivity func-
tion. Also, from this function, we derive the policy scheme for the PES-CS and pro-
pose the basic procedure for this PES-CS.
For network enforcement, we can consider the cost sensitivity based policy enforce-
ment scheme. This scheme is based on network status and cost function when a net-
work node is under a contention because of resource competition. In this section, we
define some network cost function, and then, derive cost sensitivity based policy
enforcement scheme.
For underlying network, we suppose optical networks based on GMPLS for its
control and management process. For this network, we can consider network status
such as a status under guaranteeing QoS, NS QoS and a status under contention but
guaranteeing tolerable QoS by providing alternative path, NS alt . [5] Using this net-
work status, we can define a function x (t ) , which means the binary value according
to those two statuses, NS QoS and NS alt , as follows.
Also, we can derive an operation cost model under various statuses in [5]. Using
Eq.(1) and this cost function, the total cost function when providing the service
throughout the path between the source s and the destination d is derived as follows,
Fsd (t ) = x(t ) ⋅ CQoS
sd
(t ) + (1 − x(t )) ⋅ (CQoS
sd
(t ) + Calt (t )) (2)
sd
where CQoS (t ) is a cost based on a Deterministic Effective Bandwidth (DEB) concept
in order to guaranteeing QoS by the SLA (Service Level Agreement) of a Grid service
requested by clients and Calt (t ) is a additional cost of providing a alternate path in
order to guarantee the QoS under contention situation in underlying networks [5][6].
This total cost function means the cost in order to provide the path which guarantees
the QoS.
When the data for supporting a Grid service is transmitted from source s to destina-
tion d through a network controlled by GMPLS, if the bandwidth for guaranteeing the
QoS constraint of this data is assigned, only the cost based on DEB is considered for
the total cost function. However, if the bandwidth for guaranteeing the QoS constraint
of this data can’t be assigned by the reason such as contention of resource or blocking
status, the alternate path is needed to guarantee the QoS. Thus, in this case, the total
cost function is represented by the sum of the cost based on DEB and the cost resulted
from providing the alternate path. Moreover, when it is no meaning that guarantees
the QoS because of a continuous increment of the operation cost, the total cost is
represented by the penalty cost applied differently according to the service type
204 H. Song et al.
sd
when the Fsd (t ) is given by Eq. (2) and CQoS (t ) means a cost according to DEB cost
[5], respectively. ζ F means the variance of total cost according to the variance of the
sd
cost based on DEB when traffic flows throughout the path between source s and des-
tination d.
When the sensitivity of the total cost is given by Eq. (3), we can derive the sensi-
tivity according to the network status using the total cost function, F.
∂Calt (t )
ζ Fsd = x(t ) + (1 − x(t )) ⋅ (1 + ) (4)
∂CQoS
sd
(t )
sd
When we consider the sensitivity according to CQoS (t ) , the sensitivity value of the
sd
total cost F is dominant to the variation value of both Calt (t ) . Therefore, when the
node is under a contention situation, this is, x (t ) = 0 , ζ F dominantly depends on the
sd
term, Δ = ∂C (t ) / ∂C (t ) , which represents the variation of the cost for providing the
sd
alt QoS
network operation cost point of view, so that it needs not to provide the alternate path
in this case.
In this section, we describe the PES-CS procedure and PES-CS algorithm which can
apply to Grid networks using proposed Policy based Resource Management Architec-
ture using PQRM.
Fig. 1. The basic procedure for policy enforcement scheme based on cost sensitivity with the
Policy based Resource Management using PQRM
Fig. 1 shows the basic procedure of this PES-CS for applying to Grid over GMPLS
networks. A Network Status Information Base (NSIB) is a kind of data base for col-
lected network status information and a Network Status Table (NST) is an updatable
table for selecting network resources. As shown in Fig. 1, the network status informa-
tion is saved and updated by monitoring procedure in NSIB, and NST in CS-NPEE is
created and updated by this information of the NSIB periodically. Moreover, when-
ever this NST is updated, this updated NST is enforced in that network throughout the
CS-NPEE. Moreover, as shown in a top part of Fig. 1, this update and enforcement
process interact with QoS Quorum calculated and selected by Policy based Resource
Management using PQRM. If the QoS Quorum, which matches with a specific SLA
requested by user, is selected by QoS Quorum Generator, this information is reflected
in a network enforcement process. Also, during generation of QoS Quorum in QoS
Quorum Generator, NST and network status information in NSIB are used in this
process like an upper side of Fig. 1.
In the CS-NPEE procedure, the value of NST reflects the network status informa-
tion which is represented by three network statuses. [5] That is, the value of NST,
NS , changes according to the network status, and is then updated by the proposed
ij
decision algorithm.
When it is assumed that a job with a specific SLA is requested and the contention
information in nodes is given for the update of NST, we propose the decision
206 H. Song et al.
algorithm in order to provide proper network resources by PES-CS. For these algo-
rithms or procedures, we define some factors in this section. We have the condition
factor by the number of hops, Qh , as follows:
1 if H sc − H cd ≥ 0
sd sd
Qh = ® . (5)
¯0 otherwise
sd sd
where H sc
and H are the number of passed nodes before a current nodes in this path
cd
and the number of remaining nodes after a current node in this path, respectively [5].
If the current node is closer to destination d, the value of Qh is one, otherwise, the
value of Qh is zero. Also we can obtain the other condition factor, Qδ , from the sd
F
if ζ F ≤ 1 + Δ and x (t ) = 0
sd
1
Q sd
= ® . (6)
δF
¯0 otherwise
alt DEB
the decision factor is represented by a sum of above two condition factors, the com-
bined threshold check function can then be stated as
1 if Qt =w
δ
+ wh
C ALT = ® . (7)
¯0 otherwise
Using previous parameters, we can determine the threshold check function. When
the current node between source and destination is under a contention situation, if the
node that is closer to destination d and the value of ζ F , which represents the sensitiv-
sd
ity of the total operation cost, is within the tolerable range, the combined threshold
check function C is one, so that the node makes a decision to deflect the alternate
ALT
path. Otherwise, C is zero, so that the node makes a decision to drop the data
ALT
packet. When information is obtained from NST and NSIB, the threshold check func-
tion is determined.
Finally, when the current node is under the contention situation, the node makes
a decision whether the data is to be deflected to an alternate path or dropped accord-
ing to the threshold check function, C . As mentioned in a previous part, the pro-
ALT
cedure for updating and enforcement in CS-NPEE relates to the upper part – the
Policy based Resource Management using PQRM. Fig. 2 shows the flow chart used
in CS-NPEE.
This procedure is applied by the CS-NPEE when the job with a specific SLA is re-
quested and the resource scheduling is performed by the Grid resource management
engine in the Policy based Resource Management using PQRM.
Configuration Management Policy in QoS-Constrained Grid Networks 207
1, C fr (t ) ≤ r *
Qth = ® Single Resource Allocation
¯0, otherwise
Qth = 1 ? No Table according to SLA type
where C fr (t ) means the additional of requested job is selected.
No Yes
Contention ?
Enforce determined QoS
ARQ to Network PEP.
Yes
No
Fig. 2. The Policy Enforcement Scheme based on Cost Sensitivity (PES-CS) procedure with
QoS Quorum information by the Policy based Resource Management using PQRM
scheme fluctuates widely as shown in (a) of Fig 4. Also, (b) of Fig 4 shows that most
of the sensitivity values in the case of PES-CS are constant, at 100.
1.E+00
1.E-01
1.E-03
1.E-04
1.E-05
1.E-06
1.E-07
1.E-08
0.6 0.65 0.7 0.75 0.8 0.85 0.9
Traffic Load
(a) (b)
Fig. 4. (a) The total operation cost comparison, (b) The sensitivity comparison
5 Conclusion
In this paper, we proposed policy based Grid resource management architecture based
on existing PQRM and procedure in order to manage network resource and consider
QoS guaranteed by providing efficient network enforcement scheme based on this
architecture. For this policy enforcement scheme for the modified policy based Grid
resource management architecture, we developed a modified operation cost model
according to the network status information changed by guaranteed QoS. In addition,
using the bounded range of the sensitivity of this cost, we proposed a network status
decision algorithm, and developed policy decision criteria for policy enforcement in
Grid networks by providing an alternate path. As shown in the comparison of the cost
performance between our proposed scheme and conventional schemes, our scheme is
performed under a stable state. As well, in comparing the blocking rate between our
proposed scheme and conventional schemes, ours has good performance in terms of
blocking rate. Finally, by using the bounded range of the sensitivity of the total opera-
tion cost, our proposed scheme has a reducing effect of about 24% in terms of total
operation cost.
Configuration Management Policy in QoS-Constrained Grid Networks 209
Acknowledgement
This research was supported in part by ITRC (Information Technology Res-
earch Center) and MIC (Ministry of Information and Communication) of Korean
government.
References
1. Foster, I. and Kesselman, C. (eds.). “The Grid: Blueprint for a New Computing Infrastruc-
ture”. Morgan Kaufmann, 1999
2. Foster, I. and Kesselman, C. “The Anatomy of the Grid:Enabling Scalable Virtual Organi-
zations”. Intl J. Supercomputer Applications, 2001
3. Czajkowski, K. et al. ”. Grid Information Services for Distributed Resource Sharing, 2001
4. Byung Sang Kim et al. “Policy Quorum based Resource Management Architecture in
Grids”, IJCSNS 2005, Vol 5, No 8
5. H.W. Song, S.I. Lee, C.H. Youn, “Configuration Management Policy based on DEB Cost
Model in OBS Networks”, ICAS-ICNS 2005, Oct. 2005.
6. C.H. Youn, H.W. Song, J.E. Keum, L. Zhang, B.H. Lee and E.B. Shim, “A Shared Buffer
Constrained Topology Reconfiguration Scheme in Wavelength Routed Networks”
INFORMATION 2004, Nov. 2004.
7. D. Banerjee and B. Mukherjee, “Wavelength routed Optical Networks: Linear Formula-
tion, Resource Budget Tradeoffs and a Reconfiguration Study,” IEEE/ACM Transactions
on Networking, Oct. 2000.
8. Masum Z. Hasan, et al, “Network Service Interfaces to Grid”, draft-ggf-masum-grid-
network-0, GGF Draft, May 2004.
9. Doan B. Hoang, et al, “Grid Network Services”, draft-ggf-ghph-netwerv-2, GGF Draft,
May 2005.
10. D. Simeonidou, R. Nejabati, et al, “Optical Network Infrastructure for Grid,”, draft-ggf-
ghpn-opticalnets-1, GGF Draft, Mar. 2004
11. M. Hayashi, T. Miyamoto, T. Otani, H. Tanaka, A. Takefusa, et al, ”Managing and con-
trolling GMPLS network resources for Grid applications,” OFC 2006
A Proposal of Requirement Definition Method
with
Patterns for Element / Network Management
1 Introduction
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 210–220, 2006.
c Springer-Verlag Berlin Heidelberg 2006
A Proposal of Requirement Definition Method 211
to 10-20 times the cost of rectifying errors in the requirement definition [3], it
is possible to improve the overall OSS development by improving requirement
definition. We therefore focus on improving the requirement definition process
in order to reduce the cost and time of OSS development.
We analyze the requirement definition process to enable us to improve it as
described below. In general, it is difficult to elicit complex requirement of OSSs
and create a requirement specification precisely. Consequently, an inexperienced
specifier creates low quality specifications that include errors, omissions and
lack of clarity. We believe that if the differences between knowledge possessed
by specifiers are reduced and every specifier has a high level of skill based on
that knowledge, the specifications will be of high quality. For that purpose, we
patternize some common knowledge used for creating specifications. Specifiers
refer to the patterns and can find certain solutions to problems in the requirement
definition phase.
Recently, software patterns have been used in various software development
phases, such as software design [4], and software architecture [5]. We apply such
patterns to the requirement definition phase. In fact, four patterns were intro-
duced in the requirement definition phase and their effectiveness was reported
[6]. Moreover, technologies for using patterns are required. For example, certain
development tools [7] have a function for defining patterns, and applying them to
design specifications or source codes. However, these tools don’t have functions
for the requirement definition patterns.
We propose a mechanism to facilitate pattern use in the requirement defini-
tion phases. The mechanism has functions to help specifiers select appropriate
patterns and use them unconsciously. A feature of the mechanism is the use of
roles in sequence diagrams. We aim to improve OSS development through the
requirement definition patterns and the related mechanism.
In this paper, we propose a requirement definition method that includes the
above patterns and mechanism. That is, a specifier of any skill level who uses
the method will be able to create high quality specifications easily and quickly.
First, we outline the problems related to the requirement definition phase. Then,
we detail a method for solving these problems by using patterns. After that, we
describe a mechanism that assists anyone to use the patterns easily. Finally, we
evaluate our proposed method using two case studies.
4. The specifier describes the behavior of the OSSs to realize each usecase as a
Sequence Diagram.
And, the relationships between the artifacts are as follows (Fig 1).
Business Task UC
Business Task UC
Task Task UC
It should be remembered that the specifier does not always create these dia-
grams with the above procedure. In addition, the diagrams are modified repeat-
edly after their creation and it is necessary to maintain their mutual consistency.
Client Server DB NE
1. Register (Equipment Name, Equipment ID, IP Address)
2. Retrieve (Equipment ID)
3. Send (Equipment Name)
4. Register (Equipment Name, Equipment ID)
Therefore, we think that the management data in OSSs have various roles
in a sequence diagram. For example, a role in the message 1 is “User Input”,
the role in message 2 is “Search condition of DB”, the role in the message 5 is
“Configuration data for Equipment”. We think that such roles are common in
element / network management OSSs.
In the following section we detail the mechanism to help anyone to employ
the method by using such roles
The management data roles are common to all OSSs but each OSS has certain
differences. The differences consist of which management data have roles. Figure
4 shows an example. “Equipment ID” has a role of “Configuration data for
Equipment” in System A while “Equipment No” has the role in System B.
In other words, a sequence diagram means a relationship in which management
data have roles. We named the role “Join Point” and a specifier makes a new
diagram by changing the relationship.
The procedure is as follows. We extract the roles in sequence diagram as Join
Point in advance and define the notation and structure of sequence diagram
related to Join Point. Then, we use the Join Point to generate sequence diagrams.
For example, the role “Configuration data for Equipment” generates a message
whose name is “configure ( )” from Server to NE. At this time, the generated
messages in the diagrams are defined to satisfy the notations and structures
defined by patterns. That is, specifiers can use various words (e.g. configure, set,
A Proposal of Requirement Definition Method 217
register, increase) as “to configure equipments” but patterns define only the use
“configure”. Additionally, it concerns the word as well as the structure.
In addition, a sequence diagram is generated by not only a single Join Point
but also relationships between multiple Join Points. In a simple example, a
messages order is defined by the relationships. That is, when an OSS obtains
data from a DB, there are two Join Points “search condition of DB” and “taking
data from DB”. These Join Points have a relationship whereby “search condition
of DB” comes earlier than “taking data from DB”.
Additionally, these Join Points generate some elements of a sequence diagram
and a frame of sequence diagrams can be constructed by collecting them. The
frame is not a complete sequence diagram and has some blanks that mean it is
unknown which management data have roles in Join Points (Fig.5).
Join Point
Server DB
search condition of DB Which management
1. Retrieve
data?
taking data from DB 2. Send
A frame of
sequence diagrams
Management Data Management Data
Join Point
of System A of System B
Join Point
Client Server DB NE Client Server DB NE
1. Register (Equipment Name, Equipment ID, IP Address)
2. Retrieve (Equipment ID)
Join Point 1. Register (Equipment Name, Equipment ID, IP Address)
2. Retrieve (Equipment ID)
3. Send (Equipment Name) 3. Send (Equipment Name)
4. Register (Equipment Name, Equipment ID) 4. Register (Equipment Name, Equipment ID)
– Join Points conceal the existence of sequence diagrams from specifiers be-
cause a sequence diagram is generated automatically.
– Sequence diagrams that satisfy patterns are generated.
As a result, specifiers can create sequence diagrams satisfying patterns easily
and quickly.
7 Case Study
Next, we evaluate our proposed method by applying it to the development of
two OSSs, specifically to make sequence diagrams by using Join Point. One of
the OSSs manages layer 2 networks and the other manages ATM networks. The
evaluation points are time reduction, quality of sequence diagrams such as unity,
and whether or not the method helps specifiers to make sequence diagrams.
Prior to the evaluation, we made a tool that generates sequence diagrams
by using Join Point. The tool is capable of reading relationships between man-
agement data and Join Point and generates sequence diagrams with an XML
Metadata Interchange (XMI) format.
7.1 Outline
We assume that two OSSs are developed by two specifiers who have the same
level skill and they create sequence diagrams related to the OSSs. Specifier A
uses our proposed method (patterns and Join Points) and specifier B uses the
conventional method to create the diagrams by himself. After both specifiers
finish the creation, we measure the time until the sequence diagrams are created
and their quality. Finally, we compare the time and quality, and consider whether
or not the proposed method help in the creation process.
The two specifiers create under two different conditions. One is to create new
diagrams related to OSS C. We call this case 1. The other is to change existing
diagrams to adapt to other OSS D. We call this case 2.
7.2 Result
In case 1, specifier A took 55 minutes to create five sequence diagrams while
specifier B took 93 minutes. In case 2, specifier A took 28 minutes while specifier
A Proposal of Requirement Definition Method 219
B took 55 minutes. In both cases, specifier A created diagrams more quickly than
specifier B. Moreover, the quality of the sequence diagrams created by specifier
B was poorer than that of the diagrams created by specifier A in both cases.
This was because the words used in the diagrams were varied in some places.
For example, different words were used for the same meaning, such as “access to
NE” and “get from NE”. Additionally, the diagram structures were not uniform,
for example the order of messages was different in each diagram. The two case
studies revealed that our method reduced the time and improved the quality of
diagram creation.
In addition, we wanted to evaluate whether or not the method helps specifiers
to create sequence diagrams. In order to evaluate this, we considered the spec-
ifiers’ impressions to be major indicators and we collected their comments. For
example, specifier A said that he could create and change diagrams by consid-
ering the management data that relate Join Points, and that the approach was
easier than the conventional method. This comment indicates that our method
helped him to create sequence diagrams.
Time (min) Quality Easy to create Time (min) Quality Easy to change
According to these two case studies, we got a result our method reduces time
to create sequence diagrams and improves quality of them and helps specifiers.
References
1. Frankel, D.: Model Driven Architecture. John Wiley & Sons (2003)
2. The Standish Group: The CHAOS Report (1994)
3. Boehm, B.W.: Software Engineering Economics. Prentice Hall (1981)
4. Gamma, E., Helm, R., Johnson, R., Vlissides, J.: Design Patterns: Elements of
Reusable Object-Oriented Software. Addison-Wesley (1995)
5. Buschmann, F., Meunier, R., Rohnert, H., Sommerlad, P., Stal, M.: Pattern-
Oriented Software Architecture: A System of Patterns. John Wiley & Sons (1996)
6. Hagge, L., Lappe, K.: Sharing requirements engineering experience using patterns.
IEEE Software 22(1) (2005) 24–31
7. Borland Software Corporation: Together. (http://www.borland.com/)
8. IEEE: IEEE 830 Documentation Standard for a Software Requirements Specifica-
tion. (1998)
9. Jacobson, I., Booch, G., Rumbaugh, J.: The Unified Software Development Process.
Addison-Wesley (1999)
10. ITU-T: ITU-T Recommendation M.3010 (2000)
Distributed Fault Management in WBEM-Based
Inter-AS TE for QoS Guaranteed DiffServ-over–MPLS*
1 Introduction
Recently, traffic engineering has been increasingly emphasized in end-to-end QoS
guaranteed multimedia service provisioning in next generation Internet. MPLS can
provide efficient traffic engineering by configuration of Traffic Engineering - Label
Switched Path (TE_LSP) among MPLS Label Switched Routers (LSR) [2]. In
traditional IP network, the failure in link/node has been restored after long delay from
the fault occurrence by Link State Advertisement (LSA) of routing protocol, which is
notified by flooding mechanism among adjacent routers, where each router changes
routing table to by-pass the erroneous link or node. In MPLS network, unlike the
connectionless IP network, we can achieve the high restoration performance using
protection switching function that establishes backup LSP for working LSP
dynamically or explicitly.
In 1:1 or 1+1 path protection switching scheme, the backup path should be pre-
established to provide the fast protection switching performance. However, the
backup LSP for fast restoration is not used in normal traffic transmission until any
*
This research was supported by the MIC, under the ITRC support program supervised by the
IITA.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 221 – 231, 2006.
© Springer-Verlag Berlin Heidelberg 2006
222 A. Abdurakhmanov et al.
LSP reduces the link utilization. The primary goals of fast restoration by fault
management function are (i) fast restoration of quality of service (QoS) guaranteed
differentiated path, and (ii) the guaranteed bandwidth of backup LSP at fault
occurrence.
In this paper, we design and implement distributed fault management in WBEM
based inter-AS TE for QoS guaranteed DiffServ-over-MPLS service provisioning.
We extend WBEM server with new functionality such as fault manager and SNMP
provider. SNMP provider includes SNMP trap handler module which receives SNMP
trap from the network device and fault manager to provide distributed fault
management in inter_domain DiffServ-over-MPLS network. We also propose the
design and implementation of MOF based MO for fault management in WBEM-based
inter-AS TE.
The rest of this paper is organized as follows. In section II, the related work on
WBEM-based CIM-MOF for fault management and SNMP trap notification are
briefly introduced. In section III the proposed functional architecture for WBEM-
based fault management is explained with fault protection and restoration schemes. In
section IV we explain detailed implementation of distributed fault mananagement in
WBEM-based inter-AS TE and the performance analysis. Finally, section V
concludes the paper.
2 Related Work
The DMTF has developed a core set of standards that makes up WBEM, which
includes the Common Information Model, CIM-XML, CIM Query Language, WBEM
Discovery using Service Location Protocol (SLP) and WBEM Universal Resource
Identifier (URI) mapping [3]. CIM Core Model has CIM Event Model, which
contains a mechanism to precisely identify the occurrence of the phenomenon of
interests. This model has been implemented in OpenPegasus, which is an Open
Source Software (OSS) implementation of the DMTF WBEM standard.
Fig.1 depicts OpenPegasus indication mechanism. Indication provider and
indication consumer should register on CIM server and CIM listener, respectively.
Then CIM client creates Subscription to CIM server which loads the Indication
provider. From this time, when an event occurs, the indication provider creates an
instance of the indication.
Then CIM Server checks the filters in order to determine which one allows it
through and finds those handlers which are interested in this event. The indication
consumer receives the information about indication from that handler.
The MOF for Indication hierarchy is shown in Fig. 2, and is used to describe the
type of events that can be detected. Here CIM_Indication is an abstract class and
serves as the base class for all Indication classes. This hierarchy is extensible, so new
classes can be added to capture vendor-specific properties and event types. In this
hierarchy the subclasses denote the occurrence of an action on an instance, on a class
definition and a phenomenon of interests.
Distributed Fault Management in WBEM-Based Inter-AS TE 223
Indication Handler
CIM-XML
Fig. 1. The fault indication mechanism
Fig.3 depicts WBEM based distributed fault management architecture for inter-AS
TE. In addition to traditional WBEM main components [5], there are several new
extended modules such as Fault Manager. SNMP provider is used for fault
management. The upper part of Fig.4 is WBEM Server (i.e., NMS) and lower side is a
real network element (NE) that includes nodes such as IP/MPLS router.
WBEM/CIM Server
CIMOM
Inter-domain Inter-domain
Log Data Fault Restoration Fault Restoration
CIM Client
Handler
Indication Intra-domain
Provider Fault
Handler
Handler Restoration
Handler Instance Provider
SNMP Provider
SNMP Trap Handler SNMP Manager
Net-SNMP
SNMP
SNMP SNMP
Trap
Network
NE NE NE Element
WBEM Server has three components. SNMP provider, which includes SNMP trap
handler to receive SNMP traps from SNMP agent supported network devices, and
then sends that trap to fault manager. Fault Manager, upon receiving trap, classifies
the fault, and logs to file or displays it. According to fault type it invokes local or
ingress NMS’s method provider, which includes fault restoration and fault correction
functions. According to the configuration of fault management strategy, NMS
redirects the user packet flow through the backup LSP.
Distributed Fault Management in WBEM-Based Inter-AS TE 225
Fig.4 shows the important CIM classes and associations designed for distributed fault
management in inter-domain traffic engineering.
Fig. 4. DMTF CIM based MOF classes of MPLS LSP fault management
226 A. Abdurakhmanov et al.
Classes with CIM prefix are supported by DMTF CIM, and CIM classes with QoS
prefix are newly designed MOF-based extensions for fault management in inter
domain networks. Most of the classes related to MPLS TE are derived from
CIM_EnabledLogicalElement and they have RequestStateChange() method. For
example, when MPLSTunnel down trap is received by the SNMP Trap handler, the
association class CIM_MPLSCurrenltyAssignedLSP (shown in Fig.4) is enumerated
to find the associated CIM_MPLSLSPs in that CIM_MPLSTunnel MO. The MO state
is changed to reflect the current status update. Based on the trap location and severity,
the fault manager notifies the trap to remote NMS InterASQoSProvider, where the
MPLSTunnel is originated. When the recovery is done for MOs, the notification for
recovery will also be sent to the concerned providers.
For each MPLS LSP, we can specify protection path option according to the
protection requirement of differentiated user traffic, such as 1:1, 1:N, M:N. The
backup LSP is defined by CIM_MPLSBackupLSP, which associates
CIM_MPLSTunnel and CIM_MPLSLSP. During Service level agreement between
the providers, the fault restoration method has also been negotiated and the backup
path has been configured. The backup LSP from ingress NMS to egress NMS is
usually specified to be SRLG (Shared Risk Link Group) – disjoint for the working
LSP.
for abnormal conditions on the established inter-AS TE-LSPs and also for fault
restoration. The QoS_interASNet_CMSrv is used to AS connection establishment and
gather traffic engineering information. Moreover QoS_OAMPolicy and
QoS_OAMController classes were designed to define OAM policy for distributed
fault management. All the classes are inherited from QoS_OAMService class.
The CIM_IPProtocolEndpoint MOs are associated with the link and the router’s
interfaces. For each router, which can be QoS_ASBRouter MOF (shown in Fig.6), it
should have specific identification ID attached. This ID can be used as
ingressLSRID/egressLSR ID in the MPLS Tunnels. QoS_ASBRouter contains
CIM_DeviceConnectivity_Collection and CIM_IPConnetivitySubnet MOFs which are
defined by the CIM_HostedCollection. These associations and collection classes
represent the necessary MOs for link, node and MPLS TE_LSP, which are used in
fault management.
Fault restoration procedure consists of three steps: fault detection step, fault
notification step and fault restoration step. Fault detection is performed by physical
layer detection capabilities such as Network Interface Card (NIC), MPLS signaling
such as constraint based label distribution protocol (CR-LDP) and resource
reservation protocol with traffic extensions (RSVP-TE) [8], and MPLS Operations,
Administration, and Maintenance (OAM) functions [9].
228 A. Abdurakhmanov et al.
Fault notification message using the SNMP for accurate fault localization and
diagnostics can be received from various network nodes such as router when NMS
receives SNMP trap message. However, as SNMP notification is UDP based
protocol, fault monitoring system may not receive all fault notification messages
correctly, and may not be accurate.
Fig.7 shows root cause graph in fault correlation. In this figure, we classified fault
occurrence's causes into three categories such as physical layer failure, MPLS layer
failure, and virtual private network (VPN) layer failure. The shaded box in this figure
is SNMP trap information and the others are known from each MOs. In fault
localization module, basically we used trap messages for finding upper side of failure.
For example, if path failure trap in MPLS layer is occurred, we continually analyze
whether its MO of upper layer is failure and then we have confidence of root cause of
failure.
Fig.8 shows the sequence diagram of an MPLS TE-LSP restoration. Physical layer
fault recovery is also implemented in the same manner. When SNMP handler obtains
SNMP trap from network device, it redirects this trap to fault manager’s indication
provider. Then indication provider creates instance of that trap and is redirected to
classifier function, which classifies fault type. According to fault type, it redirects to
one of the registered handlers based on filters. Classifier checks two kind of
parameters such as fault scope (intra domain or inter domain), and also type of fault
(physical or MPLS layer). In the case of physical layer fault QoS_ASBRouter MO can
be used to obtain working MPLS tunnel and accordingly failed LSP ID. In the second
case, CIM_MPLSLSP MO can be used.
As mentioned earlier, handler invokes appropriate method of CIMOM method
provider, which stores faults. Currently, four methods are used in fault re-
storation: NotifyInterDomainFaultIndication(), NotifyIntraDomainFaultIndication(),
NotifyInterDomainFaultRecovery(), and NotifyIntraDomainFaultRecovery().
The NMS of each domain is equipped with WBEM server and client functions for
inter-AS traffic engineering. When fault manager module calls fault restoration
function according to occurred fault (from local or remote NMS), it uses CIM-Client
interface. The fault management operations of fault manager module for intra-AS
traffic engineering are not open to other NMSs, but can be co-located with inter-AS
traffic engineering modules.
In this implementation the MOFs are implemented in C++ objects, and are
supported by instance providers. Currently we have implemented simple indication
provider for fault manager module to handle and process the faults, and
modified existing method provider by adding extrinsic functions, such as
NotifyInterDomainFaultIndication() or NotifyIntraDomainFaultRecovery(), which are
used for fault restoration.
Since OpenPegasus does not support SNMP manager/agent model yet, SNMP
Provider has been implemented based on Net-SNMP open source module, by
integrating it to the WBEM architecture. When instance provider calls SNMP
manager, SNMP provider acts as server, and using simple method provider interface
it provides corresponding functionality to get information from SNMP Agent. While
sending information/trap to inter-AS QoS provider or fault manager, it acts as client.
When SNMP trap handler receives trap message from SNMP agent, it triggers fault
manager by invoking notifyTrap() method of indication provider.
message will be delivered within 10-15 msec from SNMP trap handler to
fault manager. Then the fault manager immediately generates indication and
sends it to appropriate server by calling NotifyInterDomainFaultIndication() or
NotifyIntraDomainFaultIndication() functions according to fault scope. With the help
of filters and handlers, the local notification of the indication takes 60-80 msec, and in
the case of remote CIM Server it takes 350-400 msec.
600
500
400
Time [ms]
300
200
100
0
From SNMP to FM Delivery to local Delivery to remote Activation of
NMS NMS backup path
5 Conclusion
In this paper we proposed design and implementation of distributed fault management
for WBEM-based Inter-AS Traffic Engineering for QoS guaranteed DiffServ-over-
MPLS using existing CIM MOFs with hierarchical inheritance. QoS_OAMService
MOF, which designed for interASNET OAM functions, has been extended with fault
management functionalities. We also extended existing providers with SNMP
Provider, which is not supported by OpenPegasus and NET-SNMP has been
integrated into WBEM Fault Management. Also the detailed interaction scenario
among NMSs with WBEM server and client function of different AS domain
networks for fault management have been designed. Currently we are implementing
the proposed distributed fault management based OpenPegasus WBEM source code
and Net-SNMP open source tool. From initial stage performance analysis, it takes less
then 1 sec to detect, notify and restore the occurred fault.
Our future plan is to the integrate proposed distributed fault management with
Service Level Agreement and improve the performance.
Distributed Fault Management in WBEM-Based Inter-AS TE 231
References
1. Open Pegasus, http://www.openpegasus.org/.
2. Osborne Simha, Traffic Engineering with MPLS, Cisco System, 2001.
3. Web-Based Enterprise Management (WBEM) Initiative http://www.dmtf.org/standards/
wbem.
4. Douglas Mauro, Kevin Schmidt, Essential SNMP, O’Reily 2001.
5. Shanmugham S., A. Abdurakhmanov, Young-Tak Kim, “WBEM-based Inter-AS Traffic
Engineering for QoS-guaranteed DiffServ Provisioning,” proceeding of IEEE BcN 2006,
April 2006
6. Net-SNMP, http://net-snmp.sourceforge.net/.
7. Jae-Jun Jang, Young-Tak Kim, “Fault Management for GMPLS-based Optical Internet,”
Proceedings of Conference on APNOMS 2002, JeJu islands, Korea, September 2002.
8. Awduch et. Al, “RSVP-TE: Extensions to RSVP for LSP Tunnels,” IETF
9. Youngtak Kim, “MPLS OAM Functions and their Applications to Perform Monitoring,
Fault Detection and Localization,” proceeding of ACIS SERA03 (Internation Conference
on Software Engineering Research, Management and Application), SanFrancisco, May
2003.
10. Sung-Jin Lim, Design and Implementation of Differentiated Backup Path Options for Fault
Management in MPLS Networks, Master thesis, Yeungnam Univ., December 2003.
A Framework Supporting Quality of Service for
SOA-Based Applications
1 Introduction
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 232–241, 2006.
c Springer-Verlag Berlin Heidelberg 2006
A Framework Supporting Quality of Service 233
2.2 Modules
Figure 1 shows the structure of our proposed model for the framework. There
are three main blocks in this framework: interoperability control, security control
and reliability control. The purpose of interoperability control is to create SOAP
message based on service description and business rules. The SOAP message is
transferred to security control module if a secure exchange of message is required;
otherwise, the message will be transferred to reliability control module. Since the
interoperability control is supported by most popular Web Services frameworks
such as Axis or WSIF, our design does not investigate on this module. The fol-
lowing subsections present the detailed design of security and reliability modules
in our framework.
Time-to-live: time within which the client program should receive an acknowl-
edgment or result from the server provider from invoking a request.
This module also provides asynchronous invocation for client application for
better performance and loose coupling, i.e. the client should not depend on the
processing times of the Web Services. To provide these properties, this mod-
ule will be implemented in multi-thread model for communicating reliably with
server. When timeout event happen without receiving expected response, this
module re-invokes the request until the result is received.
On server side, upon receiving the request message, the module extracts Re-
liability fields (MessageID and Sequence Number ) to check for duplication. The
fault message is sent back to client if the request is duplicated, otherwise, the
module dispatches SOAP message to an appropriate handler for invoking the
requested service, starts its timer. The module sends back data to appropri-
ate module when receiving result from requested service or timeout event from
timer. Fields for managing reliability are also added to headers of the response
message before sending back to another module for returning result to the client.
To reduce the overhead in the framework, multi-thread and thread pooling mech-
anism is used in implementation both on client and server side.
On client side, multi-thread technique provides an asynchronous invocation
so that the framework can improve performance since the client could resume
its work after sending an invocation.
On server side, on receiving a request, the request message is dispatched to
the QoS Handler before invoking the appropriate remote object; new thread is
instantiated to process QoS properties as designed. The multi-thread mechanism
can be used in order to process concurrent invocations, however, it also incur
more overhead due to instantiating the threads. Thread pooling technique [15]
can be used to reduce this overhead. In this technique, threads are shared in a
pool, when needing the handlers can get a thread-worker from the pool and then
release a thread back into the pool when finishing. The pool eagerly acquires a
A Framework Supporting Quality of Service 239
Acknowledgements
This research was partly supported by Research Fund of University of Ulsan
and the Program for the Training of Graduate Students in Regional Innovation
which was conducted by the Ministry of Commerce Industry and Energy of the
Korean Government. The authors also would like to thank Ulsan Metropolitan
City and the Network-based Automation Research Center (NARC) which partly
supported this research. The authors also thank the anonymous reviewers for
their carefully reading and commenting this paper.
A Framework Supporting Quality of Service 241
References
1. Hao He, ”What Is Service-Oriented Architecture,” 2003, http://webservices.xml.
com/pub/a/ws/2003/09/30/soa.html
2. M.P. Papazoglou and D. Georgakopoulos, ”Service-oriented computing,” Commu-
nications of the ACM, Vol.46, No. 10, 2003, pp. 25-28.
3. David Booth et al., ”Web Services Architecture,” W3C Working Group Note 11,
February 2004. http://www.w3.org/TR/2004/NOTE-ws-arch-20040211/
4. Apache Software Foudation, ”Apache Axis,” http://ws.apache.org/axis/, 2006.
5. Web Services Reliability (WS-Reliability), 2004. http://docs.oasis-open.org/
wsrm/ws-reliability/v1.1/
6. SOAP Version 1.2 Part 1: Messaging Framework, http://www.w3.org/TR/soap12-
part1/, W3C Recommendation (2003).
7. XML Key Management Specification, http://www.w3.org/TR/xkms/
8. XML Encryption Syntax and Processing, http://www.w3.org/TR/xmlenc-core/
9. XML Signature, http://www.w3.org/Signature/
10. SOAP Security Extensions: Digital Signature, http://www.w3.org/TR/SOAP-
dsig/
11. Markus Volter, Michael Kircher, Uwe Zdun, ”Remoting Patterns,” John Wiley &
Sons, 2005.
12. Apache Software Foudation, ”Web Services Invocation Framework,” http://
ws.apache.org/wsif/, 2006.
13. Phung Huu Phu, Myeongjae Yi, ”A Service Management Framework for SOA-
based Interoperability Transactions,” In Proceedings of the 9th Korea-Russia Intl.
symposium on Science and Technology (KORUS2005, IEEE Press), Novosibirsk,
Russia, 2005, pp. 680–684.
14. Phung Huu Phu, ”Quality of Service for Interoperable Transactions in Service-
Oriented Architecture,” Master Thesis, University of Ulsan, South Korea, May
2006.
15. Uwe Zdun, Markus Vlter, Michael Kircher, ”Design and Implementation of an
Asynchronous Invocation Framework for Web Services,” In Proceedings of the
CWS-Europe (2003), pp. 64-78.
16. H., Santhosh K., Jen-Yao C., ”A Service Management Framework for Service-
Oriented Enterprises,” In Proceedings of the IEEE International Conference on
E-commerce Technology, California, July 06-09, 2004. pp. 181-186.
17. Y. Huang and J. Chung, ”A Web Services-based Framework for Business Integra-
tion Solutions,” Electronic Commerce Research and Applications 2 (2003), pp.15-
26.
18. G. Wang at el., ”Integrated Quality of Service (QoS) Management in Service-
Oriented Enterprise Architectures,” In Proceedings of the 8th IEEE Intl Enterprise
Distributed Object Computing Conf (EDOC 2004), California, September 20-24,
2004. pp. 21-32.
Performance Improvement Methods for
NETCONF-Based Configuration Management*
[email protected]
2
Dept. of Computer Engineering, Keimyung University, Korea
[email protected]
3
Dept. of Computer Science and Engineering, POSTECH, Korea
[email protected]
1 Introduction1
The network configuration management sets up operation values of devices that
constitute the network and collects and analyzes the values. For example, it sets up
routing parameters of routers or security values of firewalls and monitors the values.
A centralized server manages various remote devices for configuration management,
which is essential on current networks. IETF has proposed the Network Configuration
(NETCONF) [1] standard for configuration management of remote devices. The
NETCONF standard [2] assumes that the current network is composed of various
devices from diverse vendors. These standards can formulate the remote configuration
management more effectively.
Along with interoperability, the efficiency of configuration management is an
important factor to consider. Since the configuration management is carried out
against many devices, the efficiency is required thus more. For instance, when
monitoring the values of many devices, an efficient process is mandatory to achieve
* This research was supported in part by the MIC (Ministry of Information and
Communication), Korea, under the ITRC (Information Technology Research Center) support
program supervised by the IITA (Institute of Information Technology Assessment) (IITA-
2005-C1090-0501-0018) and by the Electrical and Computer Engineering Division at
POSTECH under the BK21 program of the Ministry of Education, Korea.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 242 – 252, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Performance Improvement Methods for NETCONF-Based Configuration Management 243
2 Related Work
In this section, we present implementations of XML-based network configuration
management using NETCONF by various earlier works. We also present related work
on network management performance.
The NETCONF standard is an early stage that does not have many implementations
yet. This section briefly introduces a system that is implemented by our previous
work and an open source project called EnSuite [7].
244 S.-M. Yoo, H.T. Ju, and J.W. Hong
This section discusses research work that focus on the performance of network
management using XML technologies.
Aiko Pras et al have presented Web Services for management on the performance
differences between SNMP and Web Services-based management [13]. To compare
their performances, they investigated bandwidth usage, CPU time, memory
requirements, and round trip delay. To conduct the tests, they implemented several
Web Services-based prototypes and compared their performance to various SNMP
agents. These tests showed that there is a significant difference in the bandwidth
requirements of SNMP and Web services. They concluded that SNMP is more
efficient for cases where only a single object is retrieved although Web Services-
based management is more efficient for larger number of objects. Thus, Web
Services-based management is more suitable for large scale networks.
Another interesting study on the performance of network management has been
conducted by Apostolos E. Nikolaidis et al [12]. This study mentions that the
Universal Plug and Play (UPnP) protocol undertakes the Lan configuration and
management. Due to relatively high bandwidth and the limited number of devices in
this paper, the traffic issues are of secondary importance. They examined the
unnecessary management traffic volume that may occur due to the verbose nature of
XML technology, used by the protocol. Their solution exploits some capabilities
provided by the protocol and uses real-time compression in order to effectively reduce
the management traffic, while keeping the response times at the same or even lower
levels. The solution mainly comes from the application of the Lempel-Ziv
compression algorithm, with minimal additions in the proposed DSL Forum standard.
They evaluate the performance and usability of the solution, implementing a typical
configuration example with the CPE WAN management protocol [12].
Performance Improvement Methods for NETCONF-Based Configuration Management 245
Web Browser
Manager
Message Agent
Parser Message I
Web parser N
Interfaces S
XSLT
Request T
R
Analyzer U
Request
Request M
Compress/ Compress/
rt
Tran
E
ort
Generator
Transport
coll
P
toco
Pro
ine
Decompress
e
spo
EEn
Decompress Operations N
roto
gin
Transp
ng
Block
roto
ng
Block
gin
toco
T
spo
dispatcher
Tran
ine
Pro
En
Operations Response
coll
A
E
e
dispatcher
rt
T
I
Module O
XML Parser Queue N
Repository
Repository
File RPC Operation
File RPC Operation
WSDL WSDL
XMLDB Log
XLS Configurations Configurations
The network usages and the response time have been measured to examine
NETCONF transport protocol performance. First, each transport protocol’s network
usage is examined. SOAP over HTTP has the heaviest network usages, whereas SSH
has the least, as illustrated in Figure 2 left graph. Both SOAP over HTTP and BEEP
are based on XML technology, and they both append extra messages for header
contents. Despite the fact that a message conveys only one parameter and one value,
the size of the corresponding XML message is bigger.
Next, by using each transport protocol, the response time of configuration
management is measured. In our analysis, the response time is defined as the time
interval between sending and receiving a NETCONF message. Although it was
predicted that the result of the response time would be similar to the network usage, it
has been found to differ, as Figure 2 right graph demonstrates.
Performance Improvement Methods for NETCONF-Based Configuration Management 247
The implementation methods are considered as the reason for such result. The
XCMS manager and agent use the port forwarding mechanism for SSH. It is a simple
implementation mechanism but due to the port forwarding process, some overhead
occurs. However, the actual time difference is very little in the consideration of the
unit of time and the time is round trip time.
Although the response time of the three transport protocols are similar, the network
usages are quite different. When the network usages in configuration management
have an immense effect on the network traffic, a mechanism for reducing the sizes of
NETCONF messages is needed. For solution, we used the compression method since
the repetitive nature of text patterns in typical XML message are produced by
NETCONF. Moreover, many papers have also proposed the compression mechanism
for reducing the sizes of XML. We have compressed payloads using the ZLIB library.
Fig. 3. The Network Usage and Response time of NETCONF Messages Compression
Experiments were conducted with the increasing sizes of messages to test the effect
of the compression mechanism. The response time and the network usage were
measured. Figure 3 left graph illustrates that the size difference between the original
data and the compressed data increases as the size of the data increases. Figure 3 right
graph shows that the response time depends on the network usage. Although the
process for compressing data produces overhead time, the large size data can ignore
this fact, as the number of packet fragmentations reduces.
5 RPC Layer
The NETCONF uses an RPC-based communication model. The RPC layer provides a
simple, transport-independent framing mechanism for encoding RPCs. The <rpc>
element is used to enclose a NETCONF request sent from the manager to the agent.
Next, the <rpc-reply> element encloses a NETCONF message sent in response to the
<rpc> operation on the RPC layer. The <rpc> element can only have a NETCONF
method and should have one-to-one communication with the response message. The
NETCONF provides two mechanisms for operating commands, with no pipelining
and pipelining. With no pipelining, a NETCONF manager waits for a response
248 S.-M. Yoo, H.T. Ju, and J.W. Hong
message of the previous request before sending next request message. This
mechanism has some inter-request time and low efficiency. Therefore, NETCONF
provides pipelining in order to lower the elapsed time and to improve the efficiency.
The pipelining mechanism serially sends request messages before the previous
requests have been completed, instead of waiting on a response with pipelining.
Furthermore, the NETCONF agent must send the responses only in the order the
requests were received. We investigate the effect of pipelining and propose a
mechanism for improving the efficiency on RPC layer.
We increased the number of requests containing a NETCONF command to
measure the effect of the pipelining mechanism. The command processes the ‘get-
config’ operation to obtain interface information and has a response message of
around 465bytes. The response time is measured by SSH and the other protocols have
the same result. Figure 4 demonstrates that the performance of pipelining is better
than the one of no pipelining. However, the pipelining mechanism has some risks.
The configuration management process could be destroyed if a request message is
corrupted during processing.
The pipelining mechanism does not improve the network usage performance. The
number of packets at a point of time on the network increases compared to the case
without the pipelining mechanism. The NETCONF protocol draft states that the
<rpc> element only has the NETCONF operation for its sub element. However,
several NETCONF operations related with each other are needed for a completed
configuration management. For instance, in order to complete setting the
configuration, operations for obtaining the current data as well as for setting the new
data are required.
We propose the Multi-Command mechanism for improving the efficiency of
network usage. Multi-Command is a NETCONF request message with several
commands under an <rpc> element. For example, the following three operations can
be put into a <rpc> element with Multi Command: an <edit-config> operation to
<candidate> data, <commit> operation for applying the <candidate> data to the
<running> data and <get-config> operation for checking the result of previous
operations. An agent sequentially provides the Multi-Command mechanism processes
Performance Improvement Methods for NETCONF-Based Configuration Management 249
on the request operations and creates a response. This mechanism also has some risks
that are similar to the pipelining mechanism. If any errors occur while processing a
request message, all operations of the request message are canceled by the <roll-
back> function.
The network usages and the response time have been measured to verify the effect
of the Multi-Command mechanism. We have performed a similar process to the
experiment of the pipelining mechanism and compared to three communication
mechanisms on a RPC layer. Three protocols show similar results of using the Multi-
Command mechanism, which does not have a great effect on SSH, since it adds a
small header to a payload. However, SOAP over HTTP and BEEP show larger
difference of network usages. Moreover, the communication process of the multi
command mechanism and the pipelining mechanism is similar. This mechanism also
has no inter request time but has a little additional processing time.
We have compared our results with the pipelining of HTTP. The NETCONF
pipelining has similar effect to the HTTP pipelining on response time. The requests
are buffered before transmission so that multiple HTTP requests can be sent with the
same TCP segment. As a result, HTTP pipelining does not generate unnecessary
headers and reduces the number of packets required to transmit the payload. On the
other hand, the NETCONF pipelining uses the RPC communication and transfers a
request message containing an operation at once. The NETCONF pipelining cannot
reduce the number of packets or network usages. The proposed Multi Command
mechanism uses the RPC communication and transfers a request message containing
several operations at once. Therefore, this mechanism can reduce the number of
packets and the total network usages similar to the HTTP pipelining.
6 Operation Layer
The NETCONF operations for managing configuration information of devices are
processed on an operation layer. The NETCONF protocol provides filtering methods
in selecting particular XML nodes. In particular, both of <get> and <get-config>
operations use two different filtering mechanisms. One is the subtree filtering, stated
as default by NETCONF specific and another is the XPath capability of NETCONF
protocol. These two mechanisms have their own advantages and disadvantages. The
subtree filtering is easy to comprehend but difficult to implement correctly, since the
NETCONF draft has ambiguous pats and is short of examples of accuracy on its
usage.. On the other hand, the XPath allows common XPath expressions. Thus, it is
simple to implement, but rather difficult to comprehend.
Subtree filtering and the XPath demonstrate the difference of performance.
NETCONF WG has pointed out that the XPath has a heavy memory usage and
response time for configuration management, and suggested subtree filtering as a
substitute for the XPath. The XPath requires loading the whole XML message to
DOM document before applying the request. This process efficiently creates XML
messages, but the XPath uses many memory usages regardless of complexity and
sizes. NETCONF WG has proposed that the subtree filtering is lighter than the
XPath and has the same results with the XPath. We have experimented in order to
verify these claims to confirm the difference of performance between the subtree
250 S.-M. Yoo, H.T. Ju, and J.W. Hong
filtering and the XPath. We have referenced other researches [15] of the difference
between the subtree filtering and the XPath and compared with ours. We used the
<get-config> operation and two sets of equivalent requests. The ‘Request
Message_1’ shown in Figure 5 needs to merge different processes. The result of this
request includes all interfaces whose name equals 'eth0' along with the interface
names which ‘mtu’ equals '1000'. The subtree filtering and the XPath expression
are 'Request Message_1' in Figure 5. The ‘Request Message_2’ shown in Figure 5
needs to parse simply and does not need to merge. The result of this request
includes the names of all interfaces.
Figure 5 demonstrates our experiment results of processing time and memory
usauge. The two experiments have produced rather different results. In the first
experiment, the subtree filter takes a longer time than the XPath, with the merging.
However, in the second experiment, the XPath takes a longer time than the subtree
filtering, without the merging. Moreover, these two experiments have different
XPath results. As mentioned above, the XPath [9] builds a DOM document and then
applies the XPath expression on the DOM tree. It can travel efficiently on the DOM
tree for selecting property nodes and shows more power in complicated structure. In
contrast, the subtree filtering is reliant on our implementation, which uses a
recursive top down approach for selecting nodes. The subtree filtering process is
repeated when different nodes need to be merged, which takes more time. The
XPath uses the same memory usage on both of two experiments, but it uses more
memory than the subtree filtering.
7 Concluding Remarks
In this paper, we have proposed mechanisms for effective reduction of traffic volume,
response time and computer resource usage in configuration management using
NETCONF. We have also provided an analysis of performance results by NETCONF
layers in our implementation; XCMS. In particular, we have investigated network
usages, memory requirements and response times.
The response time of transport protocol layer is hardly affected by the transport
protocols. The network usage of transport protocol layer is the sum of the
request/response message that is used to manage configuration information and the
headers of each transport protocol. Also, according to our experiments, clearly, the
response time depends on the network usages. We presented the compression
mechanism for reducing both the network usages and the response time. We also
compared our experiments to other researches of XML compression.
RPC layer provides the pipelining mechanism for more efficient configuration
management. The pipelining mechanism can reduce the total response time, but it
cannot affect the network usage. We proposed the solution of Multi-Command for
reducing total network usage, which is similar to HTTP pipelining. Our solution
reduces the response time as well as the network usages from our measurements.
Naturally, this mechanism follows a RPC communication method.
We have measured and compared the processing time and the memory usages of
the XPath and subtree filtering in our implementation. The XPATH is suitable for
processing messages since it requires fewer sources in merging the results as well as
in embedding the systems.
References
1. IETF, “Network Configuration,” http://www.ietf.org/html.charters/netconf-charter.html.
2. R. Enns, “NETCONF Configuration Protocol”, draft-ietf-netconf-prot-11, http://www.
ietf.org/internet-drafts/draft-ietf-netconf-prot-11.txt, February 2006.
3. T. Goddard, “Using the Network Configuration Protocol (NETCONF) Over the Simple
Object Access Protocol (SOAP),” draft-ietf-netconf- soap-06, http://www.ietf.org/
internet-drafts/draft-ietf-netconf-soap-06.tx t, September 16, 2005.
4. M. Wasserman, T. Goddard, “Using the NETCONF Configuration Protocol over Secure
Shell (SSH)”, http://www.ietf.org/ internet-drafts/draft-ietf-netconf-ssh-05.txt, Oct. 2005.
5. E. Lear, K. Crozier, “Using the NETCONF Protocol over Blocks Extensible Exchange
Protocol,” http://www.ietf.org/internet-drafts/draft-ietf-netconf-beep-07.txt, Sept. 2005.
6. R. Fielding, J. Gettys, J. Mogul, H. Frystyk Nielsen, L. Masinter, P. Leach and T.
Berners-Lee, “Hypertext Transfer Protocol - HTTP/1.1”, RFC 2616, IETF HTTP WG,
June 1999.
7. INRIA-LORIA, EnSuite, http://libresource.inria.fr/projects/ensuite.
8. Hyoun-Mi Choi, Mi-Jung Choi, James W. Hong, “Design and Implementation of XML-
based Configuration Management System for Distributed Systems,” Proc. of the
IEEE/IFIP NOMS 2004, Seoul, Korea, April 2004, pp. 831-844.
9. W3C, “XML Path Language (XPath) Version 2.0,” W3C Working Draft, November
2005.
10. W3C, "Web Services Description Language (WSDL) Version 1.2" July 2002.
252 S.-M. Yoo, H.T. Ju, and J.W. Hong
11. Sun-Mi Yoo, Hong-Taek Ju, James Won-Ki Hong, “Web Services Based Configuration
Management for IP Network Devices,” Proc. of the IEEE/IFIP MMNS 2005, LNCS
3754, Barcelona, Spain, Oct., 2005, pp. 254-265.
12. Apostolos E. Nikolaidis et al, “Management Traffic in Emerging Remote Configuration
Mechanisms for Residential Gateways and Home Devices,” IEEE Communications.
Magazine, Volume 43, Issue 5, May 2005, pp. 154-162.
13. A. Pras, T. Drevers, R. v.d. Meent and D. Quartel, “Comparing the Performance of
SNMP and Web Services-Based Management,” IEEE eTNSM, Vol. 1, No. 2, Dec. 2004,
pp. 1-11.
14. Mi-Jung Choi et al, “XML-based Configuration Management for IP Network Devices,”
IEEE Communications Magazine, Vol. 41, No. 7, July 2004. pp. 84-91.
15. Vincent Cridlig, et al, "A NetConf Network Management Suite:ENSUITE", Proc. of the
IEEE IPOM 2005, LNCS 3751, Barcelona, Spain, Oct., 2005, pp. 152-161.
16. Henrik Frystyk et al, "Network Performance Effects of HTTP/1.1, CSS1, and PNG,"
W3C, June 1997.
Zone-Based Clustering
for Intrusion Detection Architecture
in Ad-Hoc Networks
1 Introduction
Detecting intrusion in ad-hoc networks is harder than in regular networks. In wired or
LAN/WAN, we have a gateway where network traffic is concentrated, and an
IDS(Intrusion Detection System) can be installed there. In ad-hoc networks, we don’t
have such a convenient point. A candidate for IDS installation in ad-hoc networks
would be a node that has relatively large number of neighboring nodes located within
communication range. There should be multiple of them since we have to cover all
participating nodes in the network, and these nodes need to communicate with each
other to convey the local intrusion-related information. Finding such nodes and
allowing them to exchange information efficiently is not easy. To make the situation
worse, the nodes are mobile, and the network topology can change frequently: we
may have to repeat the hard process of setting up intrusion detection architecture
again and again.
The problem of computing efficient clusters and maintaining them in spite of
frequent changes in network topology is studied in relation with routing in ad-hoc
networks. For routing purpose, the most important parameter that determines the
efficiency of a cluster is connectivity, and the suggested techniques tend to allow
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 253 – 262, 2006.
© Springer-Verlag Berlin Heidelberg 2006
254 I.-Y. Kim, Y.-S. Kim, and K.-C. Kim
duplicate nodes (nodes belonging to more than one cluster at the same time) and to
produce many fragmented clusters (clusters with only one or two members). For
intrusion detection point of view, the critical parameter should be the number of
nodes since the cluster head is responsible for collecting security-related data from all
member nodes. Many researchers studied the above clustering problem for Ad-Hoc
IDS[1-4]. However the existing clustering techniques for IDS are simple adaptations
of ones used in Ad-Hoc routing and still have the similar problem of duplicate nodes
and fragmented cluster.
We propose a zone-based clustering technique for Ad-Hoc IDS. The aim is to
avoid duplicate nodes or fragmented clusters and to control the size of clusters to
prevent excessive load in cluster heads. Our technique clusters the given network of
nodes in two steps. In the first step, the network is divided into a set of sub-networks,
called zones. Clusters are formed within a zone in the second step; that is no cluster is
formed across zones. Zoning helps us to group geographically adjacent nodes;
clustering within this zone is a much easier problem than clustering for the whole
networks. Zoning also helps us to maintain the clusters. The replacement of a cluster
head can be handled within the corresponding zone only. The proposed algorithm has
been implemented and tested using GloMoSim simulator[5], and the result shows a
significant reduction in the cluster header’s packet processing load.
The rest of the paper is organized as follows: Section 2 examines previous studied
on Ad-Hoc IDS clustering, Section 3 explains the proposed clustering techniques in
detail, Section 4 describes experimental results, and Section 5 contains the concluding
remarks.
2 Related Work
[6] classifies IDS architectures for Ad-Hoc networks into Stand-Alone, Distributed
and Cooperative, and Hierarchical. In Stand-Alone type, all nodes have IDS, but each
node performs intrusion detection independently. A watchdog program can be
combined with Stand-Alone architecture [7]. It can monitor the packet exchange
between adjacent nodes and prevent communication with suspicious nodes.
Distributed and Cooperative type also allows all nodes to have IDS, but they should
cooperate in collecting global intrusion-related data [1].
Hierarchical architecture is one that uses clustering technique [2, 3, 4]. Fig. 1
shows an example of Hierarchical architecture. Each cluster has a cluster head, and
adjacent clusters share a gateway node as shown in Fig. 1(a). IDS’s are installed in
cluster heads as shown in Fig. 1(b). Various techniques have been developed to build
an efficient cluster and to select a cluster head. [3] suggests a technique in which the
number of neighboring nodes is computed for all nodes and one with the maximum
number becomes a cluster head. Neighboring nodes are computed based on pre-
determined number of hops: if the number of hops is 2, all nodes within 2 hops from a
node become neighboring nodes for that node. Once a cluster head is determined, the
head itself and the surrounding nodes (located within the pre-determined number of
hops from the head) form a cluster. The same process to find a cluster head and the
corresponding cluster, then, repeats for the rest of nodes. [4] suggests to build cliques
first. A clique is a set of nodes that are within 1 hop from each other and corresponds
Zone-Based Clustering for Intrusion Detection Architecture in Ad-Hoc Networks 255
to a cluster in general term. A random cluster head, then, is elected from the clique.
To protect the identity of a cluster head (for security reason), all nodes within a clique
have the same probability to be elected.
3.1 Zoning
Zoning process is distributed: seed nodes that are located at the boundary of the
network announce themselves as initiators and start to build zones at the same time.
We assume all nodes in the network know their neighbor nodes within one hop before
starting the zoning process. A node becomes a seed node if it has 1 or 2 neighbors; we
prefer lonely nodes as starting points to avoid the formation of fragmented clusters.
The seed nodes start to form zones following the algorithm shown in Fig. 2.
In the beginning all nodes are themselves zones, a one-member zone. A node can
be in several states; initially all nodes are in SLEEP state as shown in Fig. 3. A zone
must have a coordinator which determines whether to merge or not with neighbor
zones. A coordinator that actively starts the zone-merging process is called an
initiator; one that passively responds to this merging request is called a participant.
256 I.-Y. Kim, Y.-S. Kim, and K.-C. Kim
So, zoning process is a negotiation process between two coordinators that represent
two neighboring zones: one of them is an initiator, the other a participant. The state
changes of these two nodes are shown in Fig. 3.
When zoning process begins, the seed nodes send setup_init messages to their
neighbors. The seed nodes in this case act as initiators, and the neighbors become
participant. The participants send back the size of the zone they belong to. The
initiator examines the size and determines whether to merge or not. The basic
criterion is zone size after merging; if it becomes greater than some threshold, the
merging is abandoned.
Merging decision may be made for several neighbor zones at the same time. The
setup_init messages are sent to all neighboring coordinators, and if multiple responses
arrive within the same time period the initiator would consider all the corresponding
Zone-Based Clustering for Intrusion Detection Architecture in Ad-Hoc Networks 257
participants as possible candidates for merging. The response from the participant is a
setup_request message, and it contains the responder’s zone size. The initiator use this
information to give a preference to smaller participants: it is again to deter the
formation of fragmented clusters. The participant does a similar selection based on the
size of a zone. The setup_init message also contains the zone size of the initiator, and
when the participant has received a number of setup_init messages, it prefers the
initiator with the smallest zone size and sends setup_request to it. The merging process
is shown in Fig. 4(a), and the preference of the smaller zone is shown in Fig. 4(b).
When a new zone is formed through merging, a new coordinator has to be elected.
For this purpose, the old coordinator broadcasts a coordinator_election_request
message to all zone members. Each zone member responds with the neighbor node
table it has. The old coordinator counts only the neighbor nodes that do not belonging
to the current zone, and the member with the largest neighbor node number
(excluding ones within the current zone) will be elected as the new coordinator. The
node number of the new coordinator will be notified to all zone members, and this
new coordinator will start the next phase of zone merging process.
Merging process stops when all the neighbor zones are merged to the initiator’s
zone or when the size of zone has reached the maximum. When the maximum size is
reached, we have a final zone, and the state of all nodes within this zone becomes
COMPLETE. When the maximum size is not yet reached, we repeat the process of
zone merging again.
Fig. 5 shows an example of zone building process. Initially there are 26 nodes in
900m × 400m area. All of them are coordinators and form a zone by themselves as
explained in Section 3.1. When the zoning process begins, node 0, 3, 10, and 24
become the initiators because they have less than or equal to 2 neighboring nodes as
shown in Fig. 5(a). These nodes send setup_init messages to their neighbors (or
participants) and merge them. The initiators are shown in shaded circles. They are
also shown in the tables of nodes in the right side of the figure. For each table, the
shaded entry is the initiator and other entries are the participants for the initiator. The
result is shown in Fig. 5(b). In the figure, 4 new zones and the corresponding new
coordinators of them (in shaded circle) are shown.
258 I.-Y. Kim, Y.-S. Kim, and K.-C. Kim
The new coordinators tend to exist at the boundary of the zone since they have a
more number of neighbor nodes (neighbors belonging to the same zone are excluded
when computing the number of neighbor nodes). These new coordinators act as
initiators to start the next merging process. This process of merging and selecting a
new coordinator is repeated again and again until the size of all zones reach the
maximum or there is no more neighbor zone to merge. The final zones are shown in
Fig. 5(f).
3.2 Clustering
Clustering is the process of selecting a cluster head and including the neighbor nodes of
this head into the cluster. For security reason, an outsider should have no clue about who
will be and is the cluster head, and, therefore, all nodes should have an equal chance of
becoming a cluster head. Once the head is elected, the neighbor nodes are included to the
corresponding cluster until the maximum size is reached. Sharing of nodes between
clusters is prohibited, and collecting nodes from other zone is also prohibited.
Clustering is performed at each zone independently. The coordinator at each zone
at the time of completion of the zone will act as the zone manager. This manager
performs the algorithm in Fig. 6.
Basically it repeats the process of selecting a cluster head and building a cluster. In
the beginning, it sends cluster_init message to all zone members. All members
respond with a random number as shown in Fig. 7. The node which sent the biggest
random number will be chosen as a cluster head. In Fig. 7, there are 5 nodes in the
zone, and node 4 has the largest random number. Since it has neighbor node 3 and 5
as shown in figure, the first cluster will be node 3, 4, and 5 with node 4 being the
head. The clustering manager repeats the same process for the rest of nodes. The
Zone-Based Clustering for Intrusion Detection Architecture in Ad-Hoc Networks 259
second table in Fig. 7 shows node 1 and 2 become the second cluster with node 1
being cluster head.
The zones and clusters are dynamic entities. Their members may move out of them,
or new members come and join. More seriously, the zone manager or the cluster head
may move out of the region, or we may need to replace them for security reasons1. All
these events may require rebuilding of the zone or the cluster. Rebuilding is processed
in the form of merge or split – a zone or cluster whose size is below some minimum
threshold will be merged to the neighbor zone or cluster while one whose size
becomes greater than some maximum threshold will be split. The moving-out or
moving-in of a node is detected by short-term or long-term hello message. A short-
term message is a one-hop broadcast. All nodes issue short-term messages regularly
to detect the membership change. Upon receiving, it all nodes respond with their zone
and cluster identifiers. A long-term message is issued by the zone manager to all
cluster heads in the corresponding zone periodically. The cluster heads responds with
a message containing updating information such as membership change.
1
In fact, we replace the zone manager and the cluster head periodically to avoid their exposure
to persistent packet observers.
260 I.-Y. Kim, Y.-S. Kim, and K.-C. Kim
The merging of a cluster is initiated by the cluster head whose cluster size has
shrunk below some threshold. It sends a merge request to its neighbor cluster head.
When two clusters are merged successfully, a report about this change will be sent to
the zone manager by the combined cluster head. The split of a cluster is also initiated
by the corresponding cluster head. The cluster head sends a split request to the zone
manager, and the manager will start the cluster-forming algorithm in Fig. 6, but in this
case only for the members in the corresponding cluster. The merge or split of a zone
is initiated by the corresponding zone manager. Merging is essentially the same
process as the zone building process in Fig. 2.
The moving of a cluster head or a zone manager is dealt with a new election. A
new cluster head is elected when the members do not hear a short-term message from
the head for some time period. They broadcast a random number to each other, and
one issued the largest number will become the new head. Similar election is
performed at zone level when the current zone manager disappears; again this absence
of a manager is detected by the silence of the long-term message.
4 Experiments
We have implemented our zone-based clustering technique in simulated network of
mobile nodes using GloMoSim [5]. On the same network we also have implemented
two previous clustering techniques for IDS in mobile network proposed in [4] and [8].
The aims of experiments are two folds. First we compare the cluster size, and
secondly we compare the load in cluster heads. Cluster size is an important parameter
to evaluate the performance of IDS. To avoid traffic concentration on a few cluster
heads, the size should be evenly distributed among the cluster heads. Also to reduce
inter-cluster-head traffic, the number of clusters should be controlled, and for this
reason fragmented clusters (that has only one or two members) should be avoided as
much as possible. Finally to avoid unnecessary traffic monitoring, node sharing
between clusters should be prevented.
The result for cluster size comparison is shown in Table 1. In the table, WCL and
CIDS represent the clustering technique in [8] and [4] respectively, and ZIDS our
technique. Three figures are compared: number of clusters, average cluster size, and
number of single cluster. For each category, we varied the number of participating
nodes by 30, 50, and 100. The first column in the table shows the number of
clusters produced by each technique. "n" shows the number of nodes in the
network. As can be seen in the table, ZIDS is producing the least number of
clusters. The next column shows the average cluster size. This data becomes
meaningful when combined with that in the third column, which shows the number
of fragmented clusters produced by each technique. For example CIDS is producing
lots of fragmented clusters, but the average size is between those of WCL and
ZIDS. This means that CIDS is producing two kinds of clusters most of time – very
large ones and very small ones. Very large clusters will penalize the cluster head;
very small ones will increase traffic between cluster heads. On the other hand WCL
produces almost no fragmented clusters as shown in the third column. However its
average cluster size is relatively high.
Zone-Based Clustering for Intrusion Detection Architecture in Ad-Hoc Networks 261
Fig. 8 shows the numbers of packets monitored by cluster heads in CIDS and
ZIDS. The traffic was generated using CBR(Constant Bit Rate) application. We
defined a pattern file that contain 50 CBR traffic pattern, generated 10 such files, and
applied them to 50 and 100 node network respectively. As can be seen in figure, ZIDS
shows the minimum packet monitoring load: the amount of packets in ZIDS is about
32% less than that in CIDS.
50 nodes 100 nodes
300 400
cids cids
zids zids
350
250
300
200
250
observed packets
observed packets
150 200
150
100
100
50
50
0 0
p p p p p p p p p p p p p p p p p p p p
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
traffic patterns traffic patterns
5 Conclusion
In this paper, we have proposed a zone based clustering technique for intrusion
detection in Ad-Hoc network. Clustering process is essentially a distributed process
since it is hard to control all the nodes in a mobile network. However, by dividing the
network into a set of zones that contain geographically close nodes, we can control
the clustering process globally within each zone and produce more efficient clusters.
This zoning helps to produce clusters with evenly distributed size; it also facilitates
better management of clusters when the nodes move across the cluster boundary. We
have measured the performance of our technique in terms of traffic load on cluster
heads which was about 32% lighter than that in traditional clustering techniques.
References
1. Y. Zhang and W. Lee: Intrusion Detection in Wireless Ad-Hoc Networks. In: Proceedings
of the 6th Annual International Conference on Mobile Computing and Networks
(MobiCom), Boston, USA (2000)
2. D. Sterne, P. Balasubramanyam, D. Carman, B. Wilson, R. Talpade, C. Ko, R. Balupari,
C-Y. Tseng, T. Bowen, K. Levitt and J. Rowe: A General Cooperative Intrusion Detection
Architecture for MANETs. In: Proceedings of the third IEEE International Workshop on
Information Assurance (IWIA), College Park, MD, USA (2005)
262 I.-Y. Kim, Y.-S. Kim, and K.-C. Kim
1 Introduction
To deter (D)DoS attacks, technologies like Intrusion Detection System (IDS) [13],
Intrusion Prevention System (IPS) [14] and the Firewalls [15] are good solutions.
However, in reality, prevention of all attacks on the internet is nearly impossible and
the situation gets worse due to anonymous nature of IP protocol i.e. an attacker may
hide its identity if he wants to. Moreover, the routing decisions are taken on
destination addresses and none of the network unit makes sure the legitimacy of
source address. Therefore, when prevention fails, a mechanism to identify the
source(s) of the attack is needed to at least ensure accountability for these attacks and
here we need the traceback techniques.
The elements that were threatening for IPv4 networks can also be intimidating for
the future IPv6 network. To cope with IPv6 networks, we need to modify IPv4's
*
This work was supported by MIC and ITRC Project.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 263 – 272, 2006.
© Springer-Verlag Berlin Heidelberg 2006
264 S.O. Amin, C.S. Hong, and K.Y. Kim
2 Related Work
Packet Marking [1][3][4][10] algorithms are based on the idea that intermediate
routers mark packets that pass through them with their addresses or a part of their
addresses. Packets can be marked randomly with any given probability or
deterministically. The victim can reconstruct the full path with given mark packets,
even though the IP address of the attacker is spoofed. This scheme was improved in
several different ways; some of them introduced improved coding methods and
security. All of the IPv4 marking algorithms suffered by the space limitation of IPv4
header. Therefore they have to utilize encoding or fragmentation of intermediate
router’s address. The encoding of each and every packet of course degrades the
routing performance while fragmentation of address in small chunks may lead to state
explosion problem that is discussed in [7]. As a result, none of the packet marking
traceback techniques has been adapted for the practical work or implementation so
far. In our previous work, we presented a PPM algorithm for IPv6 environment which
is discussed in Section 3.
ICMP traceback [1] scheme lies under the messaging category. Every router on the
network is configured to pick a packet statistically (1 in every 20,000 packets
Tracing the True Source of an IPv6 Datagram 265
the reasons were two folds; first, the Hop-by-Hop option is processed by every router
en-route. Second, it provides the larger space to store a mark. Proposed option in Hop
by hop option header is shown in Figure 1.
Use of extension headers gave us great flexibility to pass the information to the
victim. As we marked the packet with complete address, our scheme is not vulnerable
to state explosion problem [7]. We used these marked packets to construct the reverse
routing table from victim to attackers. For this purpose, on victim side, we proposed a
data structure called Reverse Lookup Table (RLT). Following steps were taken to
complete the traceback.
1. The victim will sort the RLT by distance field; as shown in figure 2.
2. Observe the discontinuity in distance field and apply the error correction
algorithm (ECA) to find the missing nodes.
3. Finally, victim will resolve the last hop field to complete the RLT.
The resultant sorted tuples of routers can provide a complete path from Victim to
attacker.
This algorithm worked under the assumption that victim is in DDoS attack so the
number of evading packets would be sufficient to provide the information of all
routes. However, it is quite practical the victim does not have complete route
information of the attacker. For this purpose, we also introduced the Error Correction
Algorithm [17]. Marking the packet with extra 20 bytes might increases the size of
packet than PMTU, and since intermediate routers cannot do fragmentation, the
packets will be dropped. Therefore, we also proposed a modified Path MTU (PMTU)
discovery algorithm discussed in detail in [17].
Thousands of packets traverse through one router in a second and marking of every
packet, even probabilistically, may affect routing performance. Therefore, the
Tracing the True Source of an IPv6 Datagram 267
cooperation in implementing the traceback algorithm will not be tempting for ISPs.
Because it is obvious, none of the ISP provides security to other networks by
sacrificing their own customers’ satisfaction. To cope with these problems, there
should be a mechanism to minimize the burden of packet marking and initiate packet
marking only when a victim is under (D)DoS attack.
One of the ways to accomplish this is to deploy IDS on victim side and once this
IDS detects an attack it sends message to intermediate routers to initiate marking.
However, since we do not have any information of path (because we are not using
PPM here that is discussed above) we cannot send the message to desired routers to
start marking. The other option left is to multicast the message to all backbone routers
that is quite impractical due to many reason such as increase in network traffic that
may lead to network congestion. Moreover, if going along with standards, we will
have to use ICMP to send these messages and ICMP traffic is mainly filtered in many
ISPs. Therefore, there are much greater chances that these messages will be dropped
by most of the ISPs.
Another possible way is that IDSs are deployed on intermediate routers and starts
marking packets, once they detect congestion or high packet rate on any specific
interface. This scheme seems appealing by keeping in mind that most of the routers
now come with IDS functionality or we may plug-in the IDS functionality in a router
as a separate module (if this feature is present in router). The problem with this
architecture that these types of router or routers with IDS are normally deployed on
the edges of network due to the fact that adding IDS support to backbone routers will
degrade the routing performance as IDS requires high end processing to infer
something about attacks.
To mitigate the above problems we utilized the power of Policy Based Management
System [12]. Policy-based management is an administrative approach that is used to
simplify the management of a given endeavor by establishing policies to deal with
situations that are likely to occur. The description of Policy Based Management is out
of scope of this paper but it would be worthy to mention two basic building blocks of
Policy Based Management architecture i.e. Policy Decision Point (PDP) and Policy
Enforcement Point (PEP). PDP is a resource manager or policy server that is
accountable for handling events and making decisions based on those events (for
instance; at time t do x), and updating the PEP configuration appropriately. While the
PEP exists in network nodes such as hosts, routers and firewall. It enforces the
policies based on the "if condition then action" rule sets given by the PDP. Both PDP
and PEP communicates with each other through COPS (Common Open Policy
Service) that is a typical protocol [12], although DIAMETER or even SNMP may be
used.
To go with policy based management framework, of course due to standard, we
slightly modified our architecture. Instead of probabilistically marking of every
packet by intermediate routers, we maintain a list of participating edge routers (the
router closest to the sender) on PDP and placed an IDS along with traceback agent
near to the victim as shown in Fig. 3.
268 S.O. Amin, C.S. Hong, and K.Y. Kim
Once the IDS detects a (D)DoS attack on victim, it generates the request to PDP to
enforce policy which in turns, send message to all participating routers (i.e. PEP)
found in the list to initiate packet marking deterministically. Most of the IDSs detect
an attack after observing a huge traffic volume, and if we start probabilistic packet
marking after this point, we might not have large amount of marked packets to
construct the complete path. Therefore, in PBIT, we deterministically mark the
packets so one packet would be enough to get the entire path. Actually, through this
algorithm, we are not getting the entire path of an attack instead; we will be able to
get only the injection point of an attack but finding the address of an ingress point is
as good as full path traceback.
At Victim:
For traffic logging:
for every marked packet pm
if (pm.interface_addr is in RLT)
incr_packetcount(if_addr,current_time);
else{
add_in_RLT(if_addr);
set_packet_count(if_addr,1,current_time);
}
For Traceback:
If packet qm is given
If_addr=Get_ifaddrr(qm);
Else
If_addrr=max_count(RLT,time_period);
Processing Overhead: The processing can take place either at victim side or at
intermediate nodes. For an ideal traceback scheme, the processing overhead of
traceback should be minimum. Although the Figure 4 represents the traceback agent
as an integrated part but in fact it is acting as a separate component. Therefore, in
PBIT the processing overhead at intermediate nodes and victim side is almost none.
Although during traceback intermediate nodes will consume a little processing power
270 S.O. Amin, C.S. Hong, and K.Y. Kim
to mark a packet however, this kind of processing can be seen in Time To Live (TTL)
and Hop Limit calculations in IPv4 and IPv6 networks respectively. Furthermore; it is
apparent; the proposed scheme does not require any calculation of hash values or
message digests, encoding/decoding or any other computational intensive job either
on intermediate routers or at victim side.
Number of Attacking Packets: In PBIT, after (D)DoS attack detection, only one
packet is enough to complete traceback which also eliminates the path reconstruction
problem; one of the major weakness of PPM techniques.
Hash-
iTrace PPM PBIT
based
Number of
attacking Thousands 1 Thousands 1
packets
ISP
Low Fair Low Low
involvement
Network
processing Every packet Low Low Low None
overhead
During
None Low None Low
Traceback
Victim
†
processing Every packet None None None None
overhead
During
High None High Low
Traceback
Bandwidth
Every packet Low None None None
overhead
During Very
None Low None
Traceback Low
Memory
Network Low Fair None None
requirement
Victim High None High Low
Ease of
High Low Low Low
Evasion
Protection High Fair High High
Can handle
attacks other No Yes No No
then DDoS
Ease of Evasion: Refers how easily an attacker can circumvent the traceback
technique. In the case of PBIT we assume that edge routers are not compromised. For
such instances, PPM algorithm will work best due to its distributed nature.
Protection: Relates to produce the meaningful traces if some of the devices included
in traceback are undermined. PBIT is highly protective as intermediate routers don’t
participate in traceback and the single point of consideration is the router interface
closest to the attacker if this interface or a router is down then there would be no way
for an attacker to invade.
6 Conclusion
In this paper, we gave an introduction of IP traceback and a brief overview of current
IP traceback trends. These schemes were not adapted widely for IPv4 networks. One
of the main reasons was degradation in routing performance, as encoding should be
applied to pass the path information through a limited space IPv4 header.
In this paper, we discussed two Packet Marking algorithms for IPv6 network. The
extension header gave us great flexibility to pass the path information to the victim
†
Considering IDS as an external component.
272 S.O. Amin, C.S. Hong, and K.Y. Kim
and since in both of our algorithms, information of routers are not distributed in
different fragments as proposed in [3], our schemes are not affected by the state
explosion problem that is discussed in [7]. We believe that PBIT is more appealing
than PPM as it requires minimum ISP intervention and doesn’t harm the routing
performance. However, in the case of PBIT we assume that edge routers are not
compromised. For such instances, PPM algorithm will work best due to its distributed
nature.
References:
[1] Belenky, A. and Ansari, N. On IP Traceback," IEEE Communications Magazine,
Volume 41, Issue 7, July 2003
[2] S. Savage et al., "Network Support for IP Traceback," IEEE/ACM Trans. Net., vol. 9, no.
3, June 2001, pp. 226-37.
[3] Dawn X. Song and Adrian Perrig, “Advanced and authenticated marking schemes for IP
traceback,” in Proceedings IEEE Infocomm 2001, April 2001
[4] K. Park and H. Lee, "On the effectiveness of probabilistic packet marking for IP
traceback under denial of service attack," Tech. Rep. CSD-00-013, Department of
Computer Sciences, Purdue University, June 2000.
[5] A. Snoeren, C. Partridge, L. Sanchez, C. Jones, F. Tchakountio, B. Schwartz, S. Kent,
and W. Strayer. Single-packet IP traceback. ACM/IEEE Transactions on Networking,
Dec.2002.
[6] Aljifri, H. IP traceback: a new denial-of-service deterrent Security & Privacy
Magazine, IEEE , Volume: 1 , Issue: 3 , May-June 2003 Pages : 24 - 31
[7] Marcel Waldvogel, “GOSSIB vs. IP Traceback Rumors”, 18th Annual Computer
Security Applications Conference (ACSAC '02).
[8] W. Timothy Strayer, Christine E. Jones, Fabrice Tchakountio, and Regina Rosales Hain,
SPIE-IPv6: Single IPv6 Packet Traceback, Local Computer Networks, 2004. 29th
Annual IEEE International Conference on 16-18 Nov. 2004 Page(s):118 – 125.
[9] Micah Adler, “Tradeoffs in probabilistic packet marking for IP traceback,” in
Proceedings of 34th ACM Symposium on Theory of Computing (STOC), 2002.
[10] A. Belenky and N. Ansari, .On IP traceback,. IEEE Communications Magazine, vol. 41,
no. 7, July 2003.
[11] Choong Seon Hong , Pil Yong Park, Wei Jiang, " DDoS Attack Defense Architecture
Using Statistical Mechanism on Active Network Environment ", Applied Cryptography
and Network Security , pp. 47-56, June 2004
[12] A. Westerinen et al, “Terminology for Policy-Based Management”, RFC3198, IETF,
November 2001.
[13] http://en.wikipedia.org/wiki/Intrusion-detection_system
[14] http://en.wikipedia.org/wiki/Intrusion_prevention_system
[15] http://en.wikipedia.org/wiki/Firewall_%28networking%29
[16] S. Deering, R. Hinden, Internet Protocol, Version 6 (IPv6) Specification, RFC 2460,
IETF, December 1998.
[17] Syed Obaid Amin, Myung Su Kang and Choong Seon Hong, “A Lightweight IP
Traceback Mechanism on IPv6”, EUC Workshops 2006, LNCS 4097, pp. 671 – 680,
2006.
An Efficient Authentication and Simplified
Certificate Status Management for Personal
Area Networks
1 Introduction
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 273–282, 2006.
c Springer-Verlag Berlin Heidelberg 2006
274 C. Sur and K.H. Rhee
the PAN, was introduced in [3]. The personal CA is used by an ordinary user
at home or small office deployment distinguished from large scale or global CA
functions. Nevertheless, in order to use a personal PKI technology as like a
conventional PKI technology, this concept assumes that at least one device in
the PAN acts as a personal CA so as to issue certificates and provide certificate
status management to all other devices. Therefore, all the personal devices can
be equipped with certificates issued by the same CA, i.e., the personal CA, while
sharing a common root public key. As a result, mobile devices in the PAN can
establish secure and authenticated communications with each other by means of
certificates. The initialization phase of [3] extends the concept of imprinting [12]
to bootstrap all mobile devices with public key certificates. After all the mobile
devices have been imprinted with their public key certificates, mobile devices
may launch routine operations of the PAN by means of the traditional public
key signature schemes.
The personal PKI concept seems to be properly applied to PAN environment.
However, the personal PKI concept leaves at least two important challenging
problems unaddressed. The first challenging problem to think about is that: The
traditional public key signature schemes put resource-constrained mobile devices
to formidable workloads since a digital signature is a computationally complex
operation. The second challenging problem is that: To manage certificate sta-
tus information, no optimization was devised and the conventional certificate
status management schemes were considered. Consequently, to design efficient
authentication protocol and certificate status management that addresses the
aforementioned problems is a promising challenge for PAN environment.
In this paper, we propose an efficient authentication protocol that reduces
computational overheads for generating and verifying signatures on mobile de-
vices. Especially, we focus on eliminating the traditional public key operations
on mobile devices by means of one-time signature scheme, and we differentiate
it from previously proposed server-assisted computation approaches relied on
assistances of a signature server. As a result, the proposed protocol gets rid of
inherent drawbacks of server-assisted computation approaches such as problem-
atic disputes, and high computational and storage requirements on a server side.
Moreover, our protocol provides simplified certificate status management based
on hash chain technique to alleviate communication and computational costs for
checking certificate status information.
2 Preliminaries
2.1 One-Time Signatures and Fractal Merkle Tree Traveral
One-time signature (OTS for short) schemes are digital signature mechanisms
which can be used to sign, at most, one message[7]. One-time signature schemes
have the advantages that signature generation and verification are very efficient,
and further, more secure since these schemes are only based on one-way func-
tions, as opposed to trapdoor functions that are used in traditional public key
signature schemes.
An Efficient Authentication and Simplified Certificate Status Management 275
Such recursive application results in a hash chain that is generated from the
original input string:
h0 (x) = x, h1 (x), . . . , hn (x)
In most of the hash chain applications, first hn (x) is securely distributed and
then the elements of the hash chain is spent one by one starting form hn−1 (x)
and continuing until the value of x is reached.
could expire at the end of any period under the control of the certificate owner.
They intended to establish a new public key framework that exempts the CA
from testifying the validity of a certificate, once the certificate has been issued
by the CA. However, Zhou’s framework has considerable problems for practical
implementation. That is, it is an unreasonable framework to authenticate an
unidentified user based on some information submitted by the unidentified user
in exempting CA. In particular, a malicious user can always generate valid sig-
natures without any restriction. To overcome this drawback, they introduced a
new trust party called security server. However, the security server is not only a
redundant entity, but also requires an additional cost to be maintained securely.
Alternatively, we introduce control window mechanism to make Zhou’s public-
key framework above more suitable for realistic implementation.
Upon control window mechanism, CA sets the size of the control window of
the user at the certificate issuance. The user can control the status of his/her
certificate by using hash chain, and the verifier only trusts the user’s hash chain
during the control window. At the end point of the control window, the verifier
queries certificate status information to CA.
3 System Model
3.1 Design Principles and Architecture
In this section, we firstly clarify our design principles in order to efficiently
provide authentication and certificate status management among mobile devices
in PAN environment. The concerns of our design are summarized as follows:
– Eliminating Public Key Operations on Mobile Devices. Since traditional pub-
lic key signature schemes generally require computationally complex opera-
tions in terms of signature generation and even verification, they may not
even be appropriate for resource-constrained mobile devices in PAN, which
may have 8-bit or 16-bit microcontrollers running at very low CPU speeds.
Therefore, designing an authentication protocol which does not perform any
public key operations is a promising challenge in PAN environment.
– No Assistance of a Signature Server. To avoid cumbersome public key sig-
nature generations, some cryptographic protocols which depend upon a sig-
nature server were presented[1][2]. However, these approaches put a heavy
burden on the server side or, both the server and the mobile device side in
terms of high storage requirement for resolving problematic disputes. Fur-
thermore, these approaches do not eliminate public key operation on verifier
side and suffer from round-trip delay since all signing procedures are carried
out through the signature server. Consequently, it is desirable to design an
authentication protocol without assistances of the signature server.
An Efficient Authentication and Simplified Certificate Status Management 277
3.2 Notations
We use the following notations to describe the protocols:
– P CA, M : the identities of personal CA and mobile device, respectively.
– h() : a cryptographic secure one-way hash function.
– SKX : a randomly chosen secret key of the mobile device X.
i,j
– skX : the secrets of each one-time signature of the mobile device X, where
i,j
skX = h(SKX |i|j)
i is the signature number, j is the index of the secret, and | is the concatenation
of messages.
i,j i,j i,j
– pkX := h(skX ) : the commitments for each skX .
i i,1 i,t
– P LCX := h(pkX | · · · |pkX ) : the i-th public leaf commitment, which is the hash
of all the commitments of a single one-time signature.
– P KX : a public key of the mobile device X, which is the tree root of a fractal
Merkle hash tree.
– AuthP athiX : the authentication path of the i-th public leaf commitment of the
mobile device X.
278 C. Sur and K.H. Rhee
n−i
– V KX : the i-th validation key of the mobile device X. Based on a randomly
chosen secret quantity V KX from the range of h(), the mobile device X computes
0 1 n
the hash chain V KX , V KX , · · · , V KX , where
0 i
V KX = V KX , V K X = hiX (V KX ) = hX (V KX
i−1
)
n n−i
V KX constitutes X’s root validation key, V KX is X’s current validation key.
i
– SigX : the i-th one-time signature of the mobile device X.
– CertX : a certificate of the mobile device X.
4 Proposed Protocol
In this section, we present an efficient authentication protocol that provides
fast signature generation and verification without any assistance of a signature
server, and offers simplified certificate status checking by means of control win-
dow mechanism.
M , PK M , n ,VK Mn
Compute Recompute
n n
MAC k ( PCA , PK PCA , M , PK M , n , VK M ) MAC k ( PCA , PK PCA , M , PK M , n , VK M )
i
Cert M , AuthPath PCA
1. The personal CA sends its identifier and public key to a mobile device.
2. The mobile device randomly generates two secret quantities SKM and V KM .
Starting with these values, the mobile device performs the followings:
– Generates the one-time secrets/commitments pairs and the corresponding pub-
lic leaf commitments according to the total number of signatures n (Taking
into account the PAN environment, we assume that the total number of sig-
nature is less than 216 ).
– Initializes a fractal Merkle hash tree of height logn, and computes a public key
i
P KM , with the public leaf commitments values P LCM as its leaves, where
i = 1, · · · , n.
n
– Generates V KM = hn (V KM ) as the root validation key.
– Sets the signature number i = 0.
n
Then, the mobile device submits M, P KM , n, V KM to the personal CA.
An Efficient Authentication and Simplified Certificate Status Management 279
3. Both the personal CA and the mobile device carry out the following manual
authentication:
– The personal CA generates a random key k and computes a MAC as a function
n
of P CA, P KP CA , M, P KM , n, V KM by using the random key k. The MAC and
the key k are then displayed by the personal CA.
– The user now types MAC and k into the mobile device, which uses k to recom-
pute MAC value (using its stored versions of the public keys and associated
data as input).
If two values agree then the mobile device gives a success signal to the user.
Otherwise it gives a failure signal.
4. If the mobile device emits a success indication, the user instructs the personal
CA to generate a certificate. In order to generate the certificate, the personal
CA sets up a control window CW according to the system security policy
and issues the certificate signed by one-time signature for the mobile device
together with the authentication path AuthP athiP CA of the certificate.
CertM = {Ser#, M, P KM , n, V KM
n
, CW, SigPi CA },
where Ser# is a serial number.
5. The mobile device checks the followings to verify the correctness of the issued
certificate.
– Verifies the one-time signature of the personal CA on the certificate by the use
of P KP CA and AuthP athiP CA .
– Checks whether the data fields within the certificate are valid as expected.
If all the checks are successful, the protocol is now completed.
As described above, every mobile device is bootstrapped with a pair of pub-
lic/secret key and its own certificate during the initial phase. After all mobile
devices have been imprinted with their security quantities, mobile devices which
wish to sign and verify a message carry out the following signature genera-
tion/verification phase.
– Obtains CertMs and queries whether the status of CertMs is valid or not.
– Verifies the current validation key based on the root validation key in the obtained
n−i ?
certificate, i.e., hi (V KM s
n
) = V KM s
.
– If all the checks are successful, the mobile device Mv caches CertMs and sets the
current local time as starting trust time and the ending trust time based on the
control window in CertMs .
– Calculates a message digest md = h(m) and sets C = number of ’0’-bits in md ,
and then sets msg = md ||C .
– Sets SigM s
= SigMs by denoting SigM s
= {sigj }tj=1 , where t = |msg | and
updates sigj ←− h(sigj ), ∀j ∈ {j|msgj = 1}, and then calculates P LCM i
s
=
{sig1 | · · · |sigt }.
i
– Iteratively hashes P LCM s
with AuthP athiMs and compares the result to P KMs in
the certificate CertMs .
5 Evaluations
In this section, we give evaluations of the proposed protocol in terms of the se-
curity and performance points of view.
Security Evaluations. To provide secure operations, it is necessary to prove the
security of both one-time signature scheme and control window mechanism used
in the proposed protocols. Clearly, we require that message digest hash function
h() is collision-resistant. Then, it is sufficient that: if the one-way hash function
h() used for committing secrets and hash operations in the Merkle’s one-time sig-
nature scheme is a collision-resistant function which implies preimage-resistant,
no signature for a message m = m can be forged.
Regarding the security of control window mechanism, it is obvious that: to
forge the mobile device’s current validation key corresponding to the i-th one-
time signature, an adversary should compute on his own the (n − i)-th h()-
inverse of the root validation key V K n in the mobile device’s certificate, which
is computationally infeasible work.
An Efficient Authentication and Simplified Certificate Status Management 281
Table 1. gives the estimated daily communication costs according to three certifi-
cate status management schemes. If we make use of control window mechanism
instead of OCSP, then communication cost for certificate status management
can be diminished by 65%.
6 Conclusion
References
1. K. Bicakci and N. Baykal, ”Server assisted signature revisited,” Topics in Cryptol-
ogy - CT-RSA 2003, pp.143-156 March 2003.
2. X. Ding, D. Mazzocchi and G. Tsudik, ”Experimenting with Server-Aided Sig-
natures,” 2002 Network and Distributed Systems Security Symposium (NDSS’02),
February 2002.
3. C. Gehrmann, K. Nyberg and C. Mitchell, ”The personal CA - PKI for a Personal
Area Network,” Proceedings - IST Mobile & Wireless Communications Summit
2002, June 2002.
4. R. Housley, W. Ford, W. Polk and D. Solo, ”Internet X.509 public key infrastruc-
ture certificate and CRL profile,” RFC 2459, January 1999.
5. M. Jakobsson, F. Leighton, S. Micali and M. Szydlo, ”Fractal Merkel tree repre-
sentation and traversal,” Topics in Cryptology - CT-RSA 2003, pp.314-326, 2003.
6. L. Lamport, ”Password authentication with insecure communication,” Communi-
cations of the ACM, 24(11), 1981.
7. R. C. Merkle, ”A digital signatures based on a conventional encryption function,”
Advances in Cryptology - CRYPTO’87, pp.369-378, 1987.
8. R. C. Merkle, ”A certified digital signature,” Advances in Cryptology -
CRYPTO’89, pp.218-238, 1989
9. D. Naor, A. Shenhav and A. Wool, ”One-Time Signature Revisited: Have They
Become Practical?,” Cryptology ePrint Archive, Report 2005/442, 2005.
10. M. Naor and K. Nissim, ”Certificate revocation and certificate update,” The 7th
USENIX Security Symposium, January 1998.
11. M. Myers, R. Ankney, A. Malpani, S. Galperin and C. Adams, ”X.509 Internet
public key infrastructure on-line certificate status protocol (OCSP),” RFC 2560,
June 1999.
12. F. Stajano and R. Anderson, ”The resurrecting duckling: security issues for ad-
hoc wireless networks,” The 7th International Workshop on Security Protocols,
pp.172-194, 1999.
13. J. Zhou, F. Fao and R. Deng, ”An Efficient Public-Key Framework,” The 5th
International Conference on Information and Communications Security, pp.88-99,
October 2003.
A Novel Rekey Management Scheme in Digital
Broadcasting Network
1 Introduction
Digital broadcasting system utilizes CAS with hierarchic key distribution model for
access control. And 3 or 4 levels hierarchic key distribution model is a popularly used
[1]-[6]. In case of 3 levels key distribution model, control word (CW), authorization
key (AK), and master private key (MPK) are used [2], [5]. On the other hand, CW,
AK, receiving group key (RGK), and MPK are used for 4 levels key distribution
model. Note that, a disadvantage of 3 levels key distribution model compared to 4
levels one is a heavy system load in a key transmission point of view [1]-[3]. And
entitlement control message (ECM) and entitlement management message (EMM) are
used for delivering hierarchic keys [1]-[9].
CAS based on hierarchic key distribution model refreshes keys regularly and
irregularly [6]. First of all, CAS refreshes keys regularly because it provides key
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 283 – 292, 2006.
© Springer-Verlag Berlin Heidelberg 2006
284 H.-S. Koo et al.
security and efficient billing. CAS performs efficient billing by synchronizing key
refreshment period and service charging time period (CTP) [1]. However, since such
frequent key refreshment causes a big system load, a trade-off between key security
and frequent key refreshment is necessary. This regular key refreshment scheme is
called periodic entitlement management. Second of all, CAS refreshes keys
irregularly when extra key refreshment is necessary. For example, if a user wants to
terminate his/her pay service or to change his/her current entitlement to another pay
service before the entitlement is originally supposed to be expired, CAS performs
irregular key refreshment. In this circumstances, CAS generally refreshes a key
related to a channel or service group which a user wants to leave, and periodically
sends refreshed keys to all users except the one who leave his/her entitlement. This
irregular key refreshment scheme is called non-periodic or dynamic entitlement
management. Note that CAS has to send keys periodically because all digital
broadcasting standards [7]-[9] specifies one-way system as a mandatory requirement,
and two-way system as an optional one. In other words, since CAS can’t assure a
reception of refreshed keys at a host’s side in one-way system, there is no way but to
send keys periodically for reliable key transmission. Unfortunately, this mechanism
sometimes causes a big system load.
An existing solution for periodic and dynamic entitlement management has a big
flaw when it is applied to a big system with tens or hundreds pay-per-channel (PPC)
and hundreds of thousand or millions of subscribers. That is a heavy system load for
key generation and encryption [1]-[4]. Especially in case of dynamic entitlement
management, system load problem is getting more serious because a probability of
occurring extra entitlement status change events definitely will goes up compared to a
small system. This problem is what we resolved with the proposed scheme. With an
active entitlement key management proposed in this paper, CAS can handle periodic
and dynamic entitlement management with a small load and securely, even though a
system is huge.
As shown in figure 1, the active scheme has four levels key hierarchy, such as MPK,
RGK, AK, and CW. This key hierarchy model is exactly the same as the complete
scheme, but the refreshment period of AK is not CTU, but CTP. In the complete
scheme, it has to refresh AK per CTU to support dynamic entitlement management
because it is based on passive entitlement management scheme. However, our
proposed scheme broadcasts ARL to unauthorized subscribers to delete their invalid
entitlement, so we don’t need to refresh AK when a subscriber leaves his/her
entitlement.
Head-end CA server broadcasts keys via EMM after generating them. In our
scheme, there are two kinds of EMMs, which are EMMAK and EMMRGK. EMMAK and
EMMRGK are used for delivering encrypted AK, i.e., ERGK{AKs}, and RGK, i.e.,
EMPK{RGK}, respectively. These messages have periodic broadcast frequency for
reliable transmission of them. In case of EMMAK, it is broadcasted per AK
retransmission period (ARP), e.g., 0.1 ~ 15 seconds [2], and EMMRGK is transmitted
per CTP or when a subscriber subscribes new package service. Note that head-end
CA server has to generate EMMAK as many as the number of RG, and each EMMAK is
broadcasted to the corresponding RG. And there is no notation for CW and ECM in
figure 1 because they are out of the scope of this paper. Additionally, our scheme
provides the message authentication mechanism using MAC for EMMAK because it
contains ARL, and utilizes EMMAK authentication key (EAK) for the authentication
key of MAC. The details are described in the next section.
In this section, we describe the concept of ARL, the scheme of periodic transmission
of ARL over EMMAK for reliable transmission of ARL, and ARL authentication
scheme with MAC algorithm.
Second of all, if the size of ARLx is greater than the AUP/ARP, head-end CA
server organizes the table having a shape of matrix like figure 3.(b), where N is the
quotient of AUP/ARP. Note that the last row of the table might have fewer rows than
N. For example, if the value of AUP/ARP is 5 and the number of records in ARLx is
7, the ARLS transmission table for ARLx will be
288 H.-S. Koo et al.
3 Performance Analysis
Table 2 shows the number of times of key generation based on the key refreshment
frequency of the complete scheme and the active scheme with T channels, M charging
groups, and N receiving groups. Note that, the value of M is the same as the number
of days in a month [1], i.e., 30 days. First of all, the complete scheme generates AK T
times per CTU because it refreshes AK per CTU, and T×M times per CTP. In case of
RGK, the complete scheme generates RGK N+f(t) times. Here, N is the number of
columns in a row of the receiving group key matrix [1], and f(t), 1tM, indicates the
number of subscribers who leave his/her entitlement per CTU. Additionally, the
M
complete scheme generates RGK per CTP N×M+ ¦ f(t) times with the value of M
t =1
which indicates a total number of days in a month. On the other hand, the active
scheme doesn’t generate AK and RGK per CTU because it refreshes AK and RGK
per CTP, and generates AK and RGK per CTP T and N times, respectively.
Figure 4 shows the simulation results of key generation load for a system with the
assumptions like below.
1. Number of Subscribers (S) = one million, Number of Channels (T) = 100,
Number of Receiving Groups (N) = 100, and Number of Charging Groups (M) =
30.
2. The number of subscribers who leave their entitlement is 1% of S, i.e., 10,000,
and those leave events occur uniformly over a CTP. In other words, f(t) in Table
2 has a constant value of 10,000/M. at every CTU.
As shown in figure 4, the complete scheme has to generate AK and RGK about 533
(T+N+10,000/M) times per CTU, and about 16,000 (533×M) times per CTP. Besides,
it is clear that as the number of subscribers who leave their entitlement increases, the
complete scheme has to generate more keys than that the active scheme. On the other
hand, the active scheme just generates AK and RGK 200 (T+N) times per CTP. Note
that, CAS can generate them not only once in a CTP as shown in figure 4, but also
A Novel Rekey Management Scheme in Digital Broadcasting Network 289
every each day by distributing the key generation load to every day’s work. In this
situation, if we choose second approach, CAS only has to generate keys about 7 times
per CTU.
Fig. 4. The number of times of key generation Fig. 5. The number of times of key encryption
The complete
Active scheme proposed
scheme [1]
AK per CTU T None
AK per CTP T×M T
RGK per CTU N+f(t) None
M
RGK per CTP N×M+ ¦ f(t) N
t =1
Table 3 shows the number of times of key encryption of AK and RGK with S
subscribers, N receiving groups, and M charging groups. First of all, the complete
scheme has to encrypt AKs with RGK N×M times per CTU because there are N×M
packages [1] to be broadcast per CTU. When we consider it for a CTP, the complete
scheme has to encrypt AKS with RGK N×M×M times because CTP consists of M
days. In case of RGK encryption with MPK in the complete scheme, CA system
encrypts S′(t) times per CTP, and S′(t) consists of S+ f(t) + f ′(t) , 1tM, here f(t)
M
¦
t =1
M
¦
t =1
indicates the number of subscribers who leave his/her entitlement per CTU and f ′(t)
means the number of subscribers who add his/her entitlement per CTU. On the other
hand, the active scheme doesn’t need to encrypt AKs with RGK per CTU because it
doesn’t refresh AKs per CTU, and it encrypts AKs with RGK N times per CTP
because there are N receiving groups. In case of RGK encryption with MPK in the
active scheme, CA system encrypts S(t) times per CTP, and S(t) consists of S+ f ′(t ) ,
M
¦
t =1
1tM, here f ′(t) means the number of subscribers who add his/her entitlement per
CTU.
290 H.-S. Koo et al.
Figure 5 shows the simulation results of key encryption load for a system with the
assumptions like below.
1. The same as the first and second assumptions in the key generation simulation.
2. The number of subscribers who adds their entitlement is 1% of S, i.e., 10,000,
and those add events occur uniformly over a CTP. In other words, f ′(t) in Table 3
has a constant value of 10,000/M at every CTU.
With above assumptions, the complete scheme encrypts 37,000 (N×M +
(S+ f(t) + f ′(t) )/M) times per CTU, and 1,110,000 times per CTP. On the other
M
¦
t =1
M
¦
t =1
hand, the active scheme encrypts about 33,767 ((N+ S+ f ′(t ) )/M) times per CTU,
M
¦
t =1
and bout 1,013,010 times per CTP. As the simulation result shows, the active scheme
encrypts about 3,233 times less per CTU, and about 96,990 times less per CTU. Note
that if the number of subscribers who leave their entitlement increases, the complete
scheme has to encrypt more keys than the case of active scheme.
Since the active scheme additionally broadcast ARLS periodically, extra
transmission channel bandwidth consumption existed inevitably. However, the
needed amount is just negligible, and we simulate it by calculating the average bit rate
for the last AUP by varying AUP (see figure 6) to simulate it in the worst case. In
other words, the size of ARLS will be a definitely maximum at that period, and this
causes the greatest transmission channel consumption during a CTP. We assumed the
size of record in an ARL and the value of ARP as 20 bits and 15 seconds,
respectively. Note that the size of record in an ARL, 20 bits, is determined to make it
possible to identify each subscriber among one million users, and the period of ARP,
Fig. 6. Average bit rate for the last AUP by varying AUP
A Novel Rekey Management Scheme in Digital Broadcasting Network 291
4 Conclusion
In this paper, we proposed an active entitlement key management scheme for CAS on
digital TV broadcasting system. We not only introduced a novel concept of ARL for
dynamic entitlement management, but also designed key distribution model, including
ARL, based on four levels key hierarchy, ARL authentication scheme for secure
transmission of ARL, and ARL transmission table for efficient transmission
bandwidth consumption. With the proposed scheme, we can reduce key generation
and encryption load considerably compared to the complete scheme. Further, we can
manage randomly changed users entitlement status securely and efficiently with the
proposed scheme. We simulated this remarkable performance improvement by
comparing the active scheme and the complete scheme with assumptions of one
million subscribers, and one hundred PPC and receiving groups.
Acknowledgements
This research was supported by University IT Research Center Project (INHA UWB-
ITRC), Korea.
References
1. F. K. Tu, C. S. Laih, and H. H. Tung, “On key distribution management for conditional
access system on Pay-TV system,” IEEE Trans. on Consumer Electronics, Vol. 45, No. 1,
Feb. 1999, pp. 151-158.
2. H. S. Cho, and S. H. Lee, “A new key management mechanism and performance
improvement for conditional access system,” Korea Information Processing Society
Journal C, Vol. 8, No. 1, Feb. 2001, pp. 75-87.
3. W. Lee, “Key distribution and management for conditional access system on DBS,” Proc.
of International Conference on Cryptology and Information Security, 1996, pp. 82-86
4. T. Jiang, S. Zeng, and B. Lin, “Key distribution based on hierarchical access control for
conditional access system in DTV broadcast,” IEEE Trans. on Consumer Electronics, Vol.
50, No. 1, Feb. 2004, pp. 225-230.
5. ITU Rec. 810, Conditional Access Broadcasting Systems, ITU-R, Geneva, Switzerland,
1992.
292 H.-S. Koo et al.
1 Introduction
Recently, most of information systems have worked via networks. In those systems,
important information is transferred via networks. For example, users send their
personal information to a customer management system on the Web in order to enjoy
services. So, those systems need high security level in order to prevent from stealing,
leaking or tampering with customer information. Encryption is essential to protect
information through networks. Those systems also require high performance because
many customers should be able to use those systems with no stresses. However, high
security conflicts with high performance. As a system gets higher security,
performance would be lower.
Our goal is to improve performance in network systems without sacrificing
security. A network system consists of file servers and client terminals. We propose a
new approach to encrypt information in network systems. We assume that not all
information is important and need to encrypt. What is important can be defined in
advance. It would take less time to encrypt only important information. On the other
hands, we must consider keeping up security level. The key idea is encrypting only
important information and using double common-key encryption with two types of
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 293 – 301, 2006.
© Springer-Verlag Berlin Heidelberg 2006
294 K. Yanagimoto, T. Hasegawa, and M. Takano
keys i.e. static common keys and dynamic common keys in order to shorten
encrypting time and to secure systems.
The rest of this paper is organized as follows. We discuss prior work and their
limitations in section 2. We propose a new approach for system security in section 3.
Then, we describe experiments and evaluations in section 4, and discuss the approach
in section 5. Finally, we conclude with future work.
2 Prior Work
There were two approaches in system security. One is network security and the other
is database security. A lot of encrypting methods were proposed in network security.
Network security were proposed and standardized at each layer of network, such as
the application layer, the transport layer, the network layer and the data link layer [1,
2]. In those methods, payloads in each layer are encrypted. In high layers such as the
application layer, only data are encrypted. In contrast, much information is encrypted
in low layers. For example, IP addresses are encrypted in the network layer. So,
encrypting in low layer has high security but it needs more machine resources.
Encrypting payloads in low layers is secure but it would be inefficient since those
methods encrypt not only important information but also non-important information.
It also needs encrypting and decrypting at each server. Another limitation is that
performance falls when encryption keys are changed every communication in order to
secure data through network.
In database security, two popular methods were proposed. One is encrypting the all
of a database. A disadvantage of this method is low performance because stored data
need to be decrypted in data accesses. The other is encrypting fields in a database [3,
4]. This method is more efficient than encrypting all of a database. However, field
indexing is still an open problem.
3.1 Overview
We propose a new approach for security in networks and databases. Our approach is
based on double encryption to only important information. We assume that not all
information is important and need to encrypt. What is important information can be
defined in advance. It would take less time to encrypt only important information. We
must also consider keeping up security level. In order to realize it, we have two
constraints. One constraint is to store encrypted data in a database. The other
constraint is to encrypt the data so as to change the encrypted data through networks
at each communication even if the same data are sent. We propose to use two types of
common encryption keys. One is static common keys and the other is dynamic
common keys. We assume that the static common keys are common to client
terminals. We use the static common keys to store the encrypted data in a database.
We propose to use a part of network information as the dynamic common keys.
Network information is common to both a client terminal and a server, and it is
A New Encoding Approach Realizing High Security and High Performance 295
different at each communication. Since the dynamic common keys continually vary
according to network information, we use the dynamic keys to change the encrypted
data through networks at each communication. We propose double encryption using
two types of keys. First, we encrypt important information by the static common keys.
Then, we doubly encrypt the encrypted data by the dynamic common keys in order to
secure the encrypted data through network moreover. The encrypted data does not
need to be decrypted by the static common keys at the server, because we do not
encrypt the data necessary to communication such as session information and we can
store the encrypted data in a database as it is. So, performance would improve
because only decryption by the dynamic common keys is needed at the server.
Our basic idea is as follows:
1. extracting important fields from a payload in a client terminal according to
predefined information types and encrypting only the values in the important fields
by the static common keys
2. encrypting the encrypted information by the dynamic common keys moreover and
sending the doubly encrypted information to a server from the client terminal
3. decrypting the doubly encrypted information by the dynamic common key and
stored it in a database (i.e. the stored information is still encrypted by the static
common keys)
We show an example in Fig. 1. In step (1), the important data (abc) is extracted
from the payload data (abc, def, ghi) at a client terminal. In step (2), the data is
encrypted into the data (ikj) by the static common keys. In step (3), the data is
moreover encrypted into the data (pos) by the dynamic common keys. In step (4), the
payload data (pos, def, ghi) is sending to a server and the data (pos) is extracted. In
step (5), the data (pos) is decrypted into the data (ikj) by the dynamic common keys.
So, the encrypted data is stored in a database.
Client terminal
Extraction of Server
values in
Payload
important fields
data (abc,def,hig) from payload DB
(1)
Payl
Fig. 1. Overview
The encrypted information by the static common keys is doubly encrypted by the
dynamic common keys at the client terminal before sending to the server. As the
dynamic common keys, we use the identification of IP header in communication
between the client terminal and the server. The dynamic common keys are variable
since the identification of IP header is different at each communication. So the
encrypted information is continually changed even if the original information is the
same. The advantage of our method is that it does not need to exchange keys, since
the identification of IP header is known to both the client terminals and the servers.
Prior work proposed to change keys so as to change encrypted information at each
communication in order to raise security level. In order to keep the strength of the
encryption in network security, the sum of key length of a static common key and a
dynamic common key is set to the same as prior work in network security. Even
though the key length of a dynamic common key is enough short, the encrypted
information varies at each communication and the strength of security is the same
level as that of the prior work. So, another advantage of our method is that it would
not take time as long as prior work.
The doubly encrypted information is sent from the client terminal to the server and
then it is decrypted by the dynamic common keys at the server. The decrypted
information by the dynamic common keys is stored to a database as it is. The
decrypted information is secure because it is still encrypted by the static common
keys. So, system performance never falls in database access, since our approach does
not need to encrypt databases. Since encryption is done by the byte in our method, we
can access the database rightly. If the static common keys are common among another
client terminal, we can also access the database rightly by another client terminal.
4 Evaluation
We experimented with simulation based on designed systems. The system consists of
a client terminal, load balancer, a file server, a database and networks. We simulate
multi client access to a file server by using a load balancer. The system architecture of
the proposing method and prior method are shown in Fig. 3. In prior method, first, the
whole payload is encrypted at a client terminal, and then the encrypted payload is
decrypted and encrypted again at the load balancer. Finally the whole encrypted
payload is decrypted and extracted a field to data access. Note that the database is
encrypted. On the other hand, encryptions by the static common keys and the dynamic
common keys are done at a client terminal in our method. No encryptions are done at
a load balancer. Only decryption by the dynamic common keys is done at a file
server. Note that the database itself is not encrypted but important information is
encrypted. We compare the time at each component in prior method and our method.
We set the network band to 100 Mbps, and we set the average packet size to 1,200
bytes. We assume that important information exists 500 bytes from the beginning of a
payload and the size of important information is 100 byte. We used 56 bytes in DES
(Data Encryption Standard) as prior method. In our method, we set the sum of the key
length of a static key and a dynamic key to 56 bytes based on DES. 3 packets are
needed in a request and a response to a database from a client terminal respectively. A
client terminal frequently accessed to a file server up to 30 times. We used a relational
database. Data structure is shown in Table 1. 128 bit key length in AES (Advanced
Encryption Standard) was used as prior work in database security. Ten thousand
records were stored in the database.
We compared performance of each system by timing at each component, i.e. a
client terminal, a load balancer, and a file server. First, we evaluated each time in a
client terminal. The results are shown in Table 2. The time of prior method was
0.201ms and the time of our method was 0.024ms. So, the results showed that
298 K. Yanagimoto, T. Hasegawa, and M. Takano
Prior DB
method
Encryption Decryption/ Decryption/ Decryption/
Encryption Encryption Encryption
Our
method DB
5 Discussion
Our experiments showed that the transaction time at a file server is dominant and our
method successfully reduced the transaction time at a file server. We found data
access time was dominant at a file server. So, we investigated the time in data access
by varying amount of stored data in the database. We show the results in Table 3.
When the amount of stored data increased, the time increased in prior methods, while
performance does not fall in our method. Our method does not need an encrypted
database while a prior method needs an encrypted database. The result said it took
more time to access to an encrypted database when the amount of stored data
increased. So, the effect of our method does not need an encrypted database without
lowering security level. Our method also reduced the processing time at a client
terminal and a load balancer respectively.
In addition, prior method needs to exchange keys at a client terminal and a load
balancer. Our experiments showed it took extra 4.15ms to exchange keys. Our method
does not need to exchange keys because it does not need to share the keys at anytime.
300 K. Yanagimoto, T. Hasegawa, and M. Takano
Our experiments revealed that performance improved. Now we discuss the strength of
security. We compared the strength of security in prior method and our method. The
results are shown in Table 4. The encrypted block size is short in our method since the
information needs to be encrypted by the byte. It is because we can access to the
database without encrypting the database itself. However, the sum of the length of a
static encryption key and the length of a dynamic encryption key is equal to the length
of the encryption key in prior method. Our method also has an advantage of not
needing to exchange encryption keys and not needing to keep them. So, the strength
of encryption in our method would be the same level as in prior method.
5.3 Limitations
We discuss limitations of our method. One of the limitations of our method is multi
client access to a database. While the advantage of our method is not to need to keep
static encryption keys in the file server, our method forces to keep static encryption
keys in the client terminal instead. So, if multi client terminal need to access the file
server, the static common encryption keys must be shared among multi client
10
8
Extracting time (us)
0
0 300 600 900 1200
Position of important field from payload (byte)
Fig. 4. Time in extraction important fields from a payload
A New Encoding Approach Realizing High Security and High Performance 301
terminals. Our future work is to research how the static encryption key can be shared
in multi client terminals.
We did not investigate to extract an important field from a payload. As we simply
extracted an important field from the beginning of a payload, the extracting time
depended on the position of the payload. We show the extracting time according to
the position of the payload in Fig. 4. The extraction time increased in proportion to
the position of the payload. We have still room for improvement in extracting
important fields.
6 Conclusion
We proposed a double common encrypting method to only important information.
The key idea was extracting the values of important fields from a payload and
encrypting doubly the values of important fields by static common keys and dynamic
common keys. The most advantage was that our method did not need encrypt the
database itself. The experiments using a network simulator revealed that the
performance improved without falling security level. In the future, we are planning to
improve our method by tuning parameters as well as to investigate the method for
distributing the static common keys to multi client terminals when one of the client
terminals accesses to a file server for the first time.
References
1. IEEE (Institute of Electrical and Electronic Engineers), “802.1AE - Media Access Control
(MAC) Security”, http://www.ieee802. org/1/pages/802.1ae.html
2. IETF (Internet Engineering Task Force), “IETF Homepage”, http://www.ietf.org/
3. Oracle, “Transparent Data Encryption”, http://www.oracle.com/technology/oramag/
oracle/05-sep/o55security.html
4. “eCipherGate”, Hitachi System and Services, http://www.hitachi-system.co.jp/eciphergate/
(In Japanese)
GMPLS-Based VPN Service to Realize End-to-End
QoS and Resilient Paths
1 Introduction
Generalized multiprotocol label switching (GMPLS) [1] is now being applied to
multilayered networks that include IP and optical network domains. In particular, for
the core domain of the GMPLS network, optical fibers are used, and they facilitate
communication among IP domains that are connected to the core domain. In this
situation, a customer in one IP domain can communicate with another customer via a
virtual private network (VPN) on the optical core domain. From the VPN provider’s
point of view, assuring quality of service (QoS), security, and providing resilient
paths for customers is important. We apply our previously proposed hierarchically
distributed path computation elements (HDPCEs) [2] [3] to a VPN service and to
achieve these requirements.
To achieve end-to-end QoS, bandwidth assurance for each flow in an IP domain
and in a multidomain environment is indispensable. End-to-end QoS was assured
mainly by applying Differentiated Services (Diffserv) [4] [5] and Bandwidth Broker
(BB) because Diffserv aggregates many flows into a small number of Diffserv classes
and reduces the control plane burden. On the other hand, Diffserv has difficulty
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 302 – 311, 2006.
© Springer-Verlag Berlin Heidelberg 2006
GMPLS-Based VPN Service to Realize End-to-End QoS and Resilient Paths 303
2 HDPCE Architecture
First, the authorized HDPCE deployment server distributes an HDPCE software
module to HDPCE servers using the function of online Enterprise Java Bean (EJB)
[6] deployment. HDPCE modules are distributed to the interdomain HDPCE server,
which manages the VPN provider’s optical interdomain, and to the customer HDPCE
server, which is used for the customer IP domain. An example of this distribution is
shown in Fig. 1. After the distribution of the HDPCEs, each network operator
registers domain routers, optical cross connecters (OXCs), links between the routers
and OXCs, initial bandwidths, and costs of links in the distributed HDPCE. After the
registration of these links, the interdomain HDPCE establishes optical lambda paths
through optical fibers in the optical interdomain. The HDPCE can find the shortest
route for the lambda path from one domain to another using the interdomain shortest
path first (SPF) algorithm [2] [3], and bandwidths reserved between two domains are
determined by the SLA with the customer.
After the establishment of necessary lambda paths for the customer, the VPN
provider allows the customer to communicate with other IP domains via the VPN. An
example of establishing an end-to-end IP path by using HDPCEs is shown in Fig. 1.
In this example, we suppose that source IP router R1 in IP domain D_1 asks HDPCE
304 H. Matsuura and K. Takami
1 to determine the route from R1 to R11 in IP domain D_3 by a PCEP request [7]. In
this case, HDPCE 1 judges that there is no R11 in D_1 and forwards the request to the
interdomain HDPCE. The interdomain HDPCE 21 selects the appropriate lambda
path between D_1 and D_3 from its interdomain route list. Once the lambda path
whose route is R2-OXC1-R9 is chosen, the interdomain HDPCE delegates the
underlying routing to the underlying HDPCE 1 and HDPCE 3 specifying border
routers R2 and R9, respectively. The underlying HDPCEs choose the best routes in
their domains from their route lists. Finally, the optimal end-to-end route is chosen.
This route is sent back to source router R1 in the PCEP reply message [7]. Among
HDPCEs, RMI over IIOP [8] protocol is used for the interaction of EJBs.
SLA
Traffic Amount of individual traffic from customer IP domain to other IP domains via VPN.
End-to- Non- No link-state information is revealed from the customer IP domain.
end revelation
routing
Partial- Cost From the customer IP domain, BVLs, which are virtual links between
policy
VPN provider
Full- Cost From the customer IP domain, BVLs and IVLs, which are virtual links
revelation weight between BR and IR, are revealed to interdomain HDPCE.
Backup Protection In the optical interdomain, a customer-specified path is backed up by a 1:1
policy protection path which is SRLG-disjointed from the path.
Preplanned In the optical interdomain, a customer-specified path is backed up by a N:1
restoration restoration path which is SRLG-disjointed from the path.
Active In the optical interdomain, once the customer-specified path malfunctions, an
restoration alternate path is dynamically chosen.
As shown in Fig. 2, SLA terms are determined between a VPN provider and its
customer. Besides the traffic from the customer domain to other domains, there are
two main policies in SLA.
For the end-to-end routing policy, the customer can chose three options. The first
option is the “nonrevelation” policy, which reveals no virtual links (VLs) from a
customer’s IP domain. Thus, interdomain route-selection is conducted without using
link states in the customer IP domain, only the costs of interdomain lambda paths are
considered. This policy is the cheapest among the three options because the burden of
interdomain HDPCE is the lightest and the interdomain HDPCE can use effective
lambda paths for the optical domain without consideration of the customer IP domain.
The second option is the “partial-revelation” policy, which reveals border virtual
links (BVLs), which are VLs between two border routers (BRs) in the customer IP
domain. It is more expensive than the “nonrevelation” policy because an interdomain
HDPCE takes account of the VLs from the domain in addition to the interdomain
lambda paths. Applying this policy to a domain that performs the role of a hub for
other domains is very useful. For example, once a customer has three IP domains that
are connected by the VPN and one of the IP domains is mainly used for transit
between the other two domains, setting a “partial-revelation” policy in the transit
domain is a good strategy.
The third option is a “full-revelation” policy, which reveals BVLs and inner virtual
links (IVLs), which are VLs between an inner router (IR) and a BR. It is the most
expensive among the three options because the interdomain HDPCE creates all
possible routes from all registered IRs and BRs in the customer IP domain to other
domains and from other domains to all IRs and BRs in the domain.
An example in which a VPN provider has three customer IP domains that have
these three policies is shown in Fig. 3. These retrievals of VLs from underlying
HDPCEs are conducted when the interdomain HDPCE renews its interdomain route
306 H. Matsuura and K. Takami
list. An underlying HDPCE copies the corresponding route cost in its route list as the
cost of a VL when it sends VLs to the interdomain HDPCE.
In the “partial-revelation” and “full-revelation” policies, the parameter “cost
weight” should also be determined in the SLA. As shown in Eq. 1, the sum of lambda
path costs “lambdacost”s and VL costs “VLcost”s of lambda paths and VLs, which are
on the interdomain route, is the cost of the interdomain route “Routecost”.
Route cos t = x 㺌lambda cos t + 㺌y(VL cos t ) (1)
lambda route VL route
Parameter x is the cost weight of “lambdacost”s and y is the cost weight of “VLcost”s,
and the value of y is varied depending on the SLA contract with each customer.
Therefore, if cost weight y of the VLs in a customer domain is larger than those of
other IP domains, the domain is not likely to be used for the transit domain because
Routecost which passes through the domain is more expensive than other route costs.
Therefore, this high-cost-weight method is advantageous for the customer for
reserving link capacity.
There are three options in the SLA for backup policies.
The first option is protection that provides 1:1 protection for the primary path. This
option imposes a minimum delay in the case of switching paths from primary to
protection because streaming data for the primary path is copied to the protection path
in the optical domain. On the other hand, this option is expensive because 1:1
protection requires ample bandwidth in the optical network. In this option, 1:1
protection means each primary path is protected by one protection path.
The second option is preplanned restoration in which multiple primary paths share
one restoration path. Even though this option does not assure recovery from
simultaneous failures in multiple primary paths, it is cheaper than the protection
option because the restoration path is shared.
The third option is active restoration, which does not prepare the backup path
beforehand and search the restoration path when a malfunction occurs in the
customer’s path. This option is the cheapest among the three options, even though it
does not assure the backup path and has routing overhead when the path is switched
to the restoration path.
For each customer, the path backup policy can be determined on an individual-path
basis; therefore, protection, or preplanned restoration policy is likely to be applied to
important paths. We discuss these backup policies further in Section 4.
such as the link state of a domain is not kept secret with respect to other domains. To
make matters worse, optical interdomain link states of a VPN provider are revealed to
IP domains as forwarding adjacency (FA) LSPs; thus, the VPN provider cannot assure
security.
In addition, in OSPF-based legacy PCEs, Dijkstra’s shortest path first (SPF)
algorithm [11] runs whenever a PCEP setup route request comes, and it calculates the
shortest path from the source router to the destination router. The source domain PCE,
however, has an overhead of calculating the shortest path, especially if there are many
domains and links involved. In addition, in end-to-end QoS, Dijkstra’s SPF algorithm
does not check the available bandwidth of each link, so it requires the help of BB to
check whether the path has sufficient bandwidth to satisfy the QoS of each application
flow. BB, however, only checks the bandwidth of the path and does not propose the
alternative path that satisfies the bandwidth requirement.
HDPCEs resolve these problems of the OSPF protocol, and each flow-based QoS
is assured. First, HDPCEs enable the removal of link-state information flooding
among domains including the optical interdomain. This is because, as shown in Fig. 1,
each HDPCE manages the link states in its own domain and updates the link costs and
remaining bandwidth of each link after the establishment of a path in the link. VLs are
revealed by an HDPCE of an IP domain to the interdomain HCPCE, but the revelation
is restricted to the VPN provider and never revealed to other customers.
Routing delay is minimized because each HDPCE calculates the shortest routes
between possible source and destination routers/domains in its own domain and
updates these routes in its route list asynchronously with the route-setup request.
Distributed and parallel processing of HDPCEs also contributes to lighten the burden
of routing and increase the speed of the routing. In the case shown in Fig. 1, parallel
processing is conducted by HDPCE 1 and HDPCE 3 for the route selection in D_1
and D_3, respectively. If the number of VPN customers increases, dividing the
interdomain HDPCE further and preventing the concentration of the burden on one
interdomain HDPCE is also possible. For example, as shown in Fig. 1, if three new
customers join the VPN service, allocating these three customer domains under
another interdomain HDPCE and placing the inter-interdomain HDPCE on top of the
two interdomain HDPCEs is possible.
As shown in Fig. 1, an interdomain route is determined in a hierarchical manner. In
Fig. 4, we explain route-selection flow details in an HDPCE. As shown in Step 3, an
HDPCE chooses the route on a route-cost basis and on an available bandwidth basis
from the routes in its route list. After the determination of the route in the domain, the
HDPCE updates the remaining bandwidth and cost of the links on the route, as shown
in Step 4. Therefore, the probability of finding the path that satisfies the bandwidth
requirement becomes greater compared with that of the combination of the OSPF
protocol and BB.
The SLA of the end-to-end routing policy, “full revelation policy,” has the highest
probability of accommodating the most paths that satisfy the required bandwidth in
the customer domain. This is because the cost of a VL is determined by the least-
loaded (LL) algorithm [12], where the cost is defined as the inverse of the remaining
bandwidth. Therefore, VLs that have less remaining bandwidth are not likely to be
used. On the other hand, links in a customer domain that has a “nonrevelation policy”
are likely to be used without consideration of remaining bandwidth; thus,
308 H. Matsuura and K. Takami
lower HDPCE
YES
Step 3: Choose the cheapest route which satisfies the bandwidth required.
Step 4: Update the remained bandwidth and costs of links on the route.
QoS-assured end-to-end routes are less likely to be created than those of other
customer domains.
with other group members. In legacy OSPF-based routing, however, selecting an SRLG
disjointed from the primary path is difficult because Dijkstra’s SPF algorithm does not
take SRLG information into account and only considers link costs [14].
As is shown in Fig. 5, interdomain HDPCE establishes lambda paths based on the
required traffic among IP customer domains and the interdomain route-costs. Each
lambda path is related to its component SRLG sequence and its cost in the lambda
path list. Sometimes a lambda path passes through one or more than one IP domains,
which are transit domains. lambdacost is defined as
lambdacos t ( n, B, Į, ȕ ) = Įȕ / B + (100 - Į )n , (2)
where n indicates the number of route components of the lambda path such as OXCs
and optical fibers, B indicates the available bandwidth of the lambda path, Į is used to
weight of the inverse of B against n within the range between 0 and 100. ȕ is used for
compensating between the inverse of B and n before the IP paths are accommodated
in the lambda path. On the other hand, Į is flexibly changed depending on the
conditions of the applied interdomain. For example, if the number of disjointed SRLG
pairs between any two IP domains becomes limited, setting Į to a small value for
effective accommodation of the IP paths in the lambda paths is better. That is because
selecting a lambda path that has a small number of route components for a primary IP
path helps to select a protection IP path, which is SRLG-disjointed from the primary
path. On the other hand, for large networks that have many alternative disjointed
SRLG routes between any two packet domains, setting Į to a large value for effective
accommodation of the IP paths in the lambda paths is better. That is because available
bandwidth B is weighted more and balanced use of lambda paths is conducted.
A flowchart of how an interdomain HDPCE provides three different backup
methods for individual IP paths is shown in Fig. 6. Step 1 shows the flow for the
selection of a primary IP path, but if some optical fiber of the primary IP path has a
exception NO Step 5
NO exception Is there the protection path ?
Step 9
Is there the restoration group ? YES
YES
malfunction, Step 1 is also used to select an active restoration path excluding the
malfunctioning SRLG. The protection IP path for the primary IP path is selected in
Step 4 in which a disjointed SRLG lambda path is chosen from the lambda path list to
accommodate the protection IP path.
In addition to the lambda path list, the interdomain HDPCE maintains a list of
restoration groups. A restoration group consists of SRLG-disjointed primary lambda
paths and their shared restoration lambda path. In Step 6, the interdomain HDPCE
searches the restoration group that includes the same lambda path in which the
primary IP path is accommodated. If such a group exists, the restoration lambda path
in the restoration group is used for the backup of primary IP path. If there is no such
group, the primary lambda path is added to another restoration group, or a new
restoration group is created in Step 8.
The GMPLS network shown in Fig. 7(a) is used for the evaluation of two proposed
backup methods: protection and preplanned restoration methods. The network has an
optical interdomain managed by an interdomain HDPCE and two underlying IP
domains, namely D_1 and D_2, managed by two corresponding HDPCEs. First, we
set up lambda paths that have ten units of bandwidth from each BR in D_1 to each BR
in D_2. Therefore, after the lambda path setups, there are nine lambda paths in the
network, as shown in Fig. 7(a). After setting up lambda paths in the GMPLS
networks, we set up IP paths that require one unit of bandwidth from the underlying
inner router, R1, to the other inner router, R2, using two different backup methods.
In this evaluation, ȕ in Eq. 2 is set to 10. As shown in the figure, preplanned
restoration backup accommodates 20 – 80% more primary IP paths than protection
backup at the expense of vulnerability against simultaneous malfunctions. The larger
the value of Į is, the lower the accommodation rate of primary IP paths is in the
protection method, because this network, which has only 9 lambda paths, is a
relatively small network. Preplanned restoration, however, has a consistent
accommodation rate regardless of the value of Į; thus, the network operator of a VPN
GMPLS-Based VPN Service to Realize End-to-End QoS and Resilient Paths 311
provider does not have to adjust the value of Į. The active restoration method
accommodates the most primary paths among the three methods, though the SRLG-
disjointed backup path is not necessarily assured.
5 Conclusion
We described an HDPCE-based VPN service that provides IP domain customers with
end-to-end QoS assurance, security, and three flexibly chosen backup methods.
First, HDPCEs provide individual-flow-based multidomain routing, which assures
the required bandwidth of each flow. Second, link-state information of each customer
is never revealed to other customers, so security is assured. Third, three backup
methods: protection, preplanned restoration, and active restoration are flexibly chosen
depending on the priority and backup strategy of each path of the customer.
References
1. A. Bonerjee, J. Drake, J. P. Lang, and B. Turner, “Generalized multiprotocol label
switching: an overview of routing and management enhancements,” IEEE Commun. Mag.,
vol. 39, no. 1, pp. 144-150, Jan. 2001.
2. H. Matsuura, T. Murakami, and K. Takami, “An Interdomain Path Computation Server for
GMPLS Networks,” IEICE Trans. Commun., Vol. E88-B, No. 8, pp. 3329-3342, August
2005.
3. H. Matsuura, N. Morita, T. Murakami, and K. Takami, “Hierarchically Distributed PCE
for GMPLS Multilayered Networks,” IEEE Globecom 2005, St. Louis, Missouri, USA,
Nov. 2005.
4. R. Bless, “Towards Scalable Management of QoS-based End-to-End Services,” IEEE/IFIP
Network Operations and Management Symposium, April 2004.
5. Performance Evaluation for a DiffServ Network’s PHBs EF, AF, BE and Scavenger,”
IFIP/IEEE International Symposium on Integrated Network Management, May 2005.
6. Sun Microsystems, “Enterprise JavaBeansTM Specification Version 2.1,” August 2002.
7. J. P. Vasseur, J. L. Roux, A. Ayyangar, E. Oki, A. Atlas, and A. Dolganow, “Path
Computation Element (PCE) communication Protocol (PCEP) Version 1,” IETF Internet
Draft, draft-vasseur-pce-pcep-02.txt, Sep. 2005.
8. Object Management Group, “CORBA V2.3.1,” Oct. 1999.
9. N. Bitar, R. Zhang, and K. Kumaki, “Intar-AS PCE Requirements,” IETF Internet Draft,
draft-bitar-zhang-inter-AS-PCE-req-011.txt, Oct. 2005.
10. J. L. Roux, “PCE Communication Protocol (PCECP) specific requirements for Inter-Area
(G)MPLS Traffic Engineering,” draft-leroux-pce-pcep-interarea-reqs-00.txt, Oct. 2005.
11. E. W. Dijkstra, “A Note on Two Problems in Connexion with Graphs,” Numerische
Mathemakik, 1, pp. 299-271, 1959.
12. Q. Ma and P. Steenkiste, “On Path Selection for Traffic with Bandwidth Guarantees,” In
Proceedings of IEEE International Conference on Network Protocols, October 1997.
13. D. Xu, Y. Xiong, C.Qiao, and G. Li, “Trap avoidance and protection schemes in networks
with shared risk link groups,” in IEEE Journal of Lightwave Technology, Special issue on
Optical Network, Nov. 2003.
14. E. Oki et al., “A Disjoint Path Selection Scheme with SRLG in GMPLS networks,” Proc.
of IEEE HPSR’2002, 88-92, May 2002.
WBEM-Based SLA Management Across Multi-domain
Networks for QoS-Guaranteed DiffServ-over-MPLS
Provisioning*
1 Introduction
In order to provide on demand QoS-guaranteed DiffServ-over-MPLS service
provisioning across multiple autonomous system (AS) domain networks, the network
management system (NMS) should be able to collect the information of inter-AS
connectivity and available network resources among autonomous system boundary
router (ASBR) ports, and fault restoration capability of the traffic engineering label
switched path (TE-LSP) [1]. For this end-to-end QoS-guaranteed DiffServ
provisioning management, NMS must establish service level agreement (SLA) with
other providers and with the customers. The NMS needs to interact with other NMSs
of different network providers for the QoS-guaranteed TE-LSP connection
establishment and this interaction is required to be platform and network resource
*
This research was supported by the MIC, under the ITRC support program supervised by
the IITA.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 312 – 321, 2006.
© Springer-Verlag Berlin Heidelberg 2006
WBEM-Based SLA Management Across Multi-domain Networks 313
2 Background
WBEM is a platform and resource independent DMTF standard that defines both a
common model and protocol for monitoring and controlling resources from diverse
sources. DMTF defines the CIM that defines the management information for system,
networks, application and services, etc. CIM includes the description of a meta-
language for describing data (i.e., CIM specification) and a description of the
resources to be managed (i.e., CIM schema and vendor extensions). CIM-XML is a
DMTF standard for WBEM communication protocol that uses the CIM specification
and CIM schema for the representation of managed resources, defines the encoding of
CIM data and operations into XML, and uses HTTP transport protocol for exchanging
CIM-XML encoded requests and responses [2]. CIM MOF is the language defined by
the DMTF for describing classes and instances.
The SLS is used to denote the technical characteristics of the service offered in the
context of an SLA. The service technical characteristics refer to the provisioning aspects
of the service, such as request, activation and delivery aspects from network perspectives.
Non-technical service provisioning aspects such as billing and payment aspects, are not
part of the SLS; they are part of the overall SLA. SLS forms the basis of the agreements
between providers for traffic exchange in the Internet. SLS include SLS identification,
scope (geographical/topological region), flow identification, traffic conformance, excess
treatment, performance guarantees, schedule, etc. Our work draws from the SLS template
specification work of the MESCAL SLS specification [6 - 8].
QoS-based services are offered on the basis of the SLAs, which set the terms and
conditions on behalf of both providers and customers in providing and requested
services, respectively. Before committing any connection establishment, SLA
negotiation should have been made between the providers. There is a performance
monitoring part, for each negotiated QoS parameters (such as delay, jitter, packet loss,
and throughput), which is composed of (i) Measurement Period, (ii) Reporting, and
(ii) Notification Threshold [7]. The performance management would perform
monitoring the QoS parameters that had been negotiated in the SLA so as to check
that whether SLA is violated or not.
Fig.1 depicts the functional block of WBEM-based SLA management for inter-AS
TE-LSP connection establishment. The OpenSLP server acts as a dedicated DA
where all NMSs register and discover the service contact points using service location
protocol (SLP).
WBEM-Based SLA Management Across Multi-domain Networks 315
The interaction among the NMSs, DA and CNM (customer network manager) for SLA
negotiation is shown in the Fig. 2 Before the interaction among NMSs is done, the NMS
need to perform initialization of its providers. During NMS provider initialization, it
tries to register itself to the DA, and to discover other NMSs that participate in inter-AS
TE from the DA using SLP protocol. Each NMS tries to gather the resource details from
other NMSs to make inter-AS topology configuration. After the topology configuration
is accomplished, SLA negotiation for inter-AS TE can be handled.
316 J.-C. Seo et al.
Fig. 2. Interaction sequence diagram for SLA negotiation in inter domain networks
Fig.3 depicts unified modeling language (UML) representation of CIM classes and
associations designed for SLA negotiation and connection establishment in inter-AS
TE. Some classes are designed as abstract classes and other classes inherit from the
WBEM-Based SLA Management Across Multi-domain Networks 317
The inter-AS QoS provider which is shown in Fig.1, acts as core part of our WBEM
architecture. This provider includes instance provider, association provider and
method provider functions for inter-AS TE (traffic engineering). When WBEM server
is loaded, the Inter-AS QoS Provider gets service location list of other providers from
SLP provider and configures the inter-AS connectivity topology. When SLA is
negotiated by the customer (or other providers) with requested parameters, it
318 J.-C. Seo et al.
computes CSPF routing for the requested destination based on the up-to-date inter-AS
topology details. When NMS needs to establish TE-LSP connection setup and backup
path, the method provider starts to establish connection by interacting with SNMP
provider or CLI Provider in the WBEM server. All the provider modules are
implemented using C++.
SLP provider in NMS consists of a service agent, user agent and CIM client. When
the WBEM server is loaded, the SLP provider registers Inter-AS supported WBEM
Pegasus service to statically registered DA. SLP provider periodically sends unicast
service request message to DA to get the service location list of registered NMSs.
Using CIM SLP Template instances, the SLP provider stores the details of the
registered NMSs.
Fig.4 shows the SLP provider operation for service register and discovery. In our
implementation, DA address is statically configured in the NMS’s SLP provider. The
DA is implemented in Linux platform using OpenSLP[3]. The UA functionalities are
provided by the OpenPegasus 2.5.1 client API.
SLA provider is used to negotiate the SLA between the providers and the customers.
Before establishing any data transfer for on-demand connection request, the providers
and customers negotiate each other for the mutual agreement on the proposed
parameters in the SLS. The SLA negotiation between the providers and customers are
implemented using the WBEM CIM model. Fig.2 explains the SLA negotiation
process between the customers and the providers. The client initiates the negotiation
by SLA_Open() to the provider. The SLA provider accepts the negotiation session by
WBEM-Based SLA Management Across Multi-domain Networks 319
SLA_Accept(). When the client sends the SLA_Request() with the required
parameters for on-demand TE-LSP connection provisioning, the SLP provider checks
whether the managed domain has enough resource for the new connection. This is
done by calling getTEInfo(), which in-turn calls getUptodateLocalStateInfo() method
of Inter-AS QoS provider. The current network resources are checked and if the
managed domain has enough resource, the service level negotiator will initiate the
service level negotiations to other SLP providers who are in the route to the
destination. If there is not sufficient resources available, the SLP provider can re-
negotiate to the requested parties by calling SLA_Decision() request with re-
negotiation option. Once the ingress NMS receives successful SLA_Decision() from
all the NMSs, they can enforce the NE configuration by using SNMP provider or CLI
provider. The ingress NMS calls the customer with the decision by SLA_Desicion().
The decision could be either accept/reject or re-negotiate. The customer can send the
report by SLA_Report() to the ingress-NMS, which conveys that the customers
decision to the provider about the negotiation.
Fig. 5 shows time taken by the SLA negotiation process with five NMS. When the
ingress NMS gets a request for SLA negotiation for TE-LSP connection
establishment, it checks itself with getTEInfo(). The request will perform association
traversals and instance enumerations on the repository. The graph depicts the time
taken for enumerating the instances the CIMOM repository.
1.9
Time (sec)
1.8 1.81
1.79
1.77 1.769 1.77 1.77
1.7
1.6
1.5
1000 5000 10000 20000 50000 100000
Number of Instances
From the graph, it is clear that there is not much variance in time taken and the
time taken is irrespective to the number of instances enumerated. On the average, it
takes almost 1.78 seconds. It shows the better scalability for the number of instances
to be enumerated. The ingress SLA provider will interact with other four NMSs SLA
providers sequentially and the cumulative time taken is around 6~8 sec for end-to-end
negotiation.
Fig. 6 shows the complete time taken by each functions of WBEM architecture
involved for SLA negotiation for TE-LSP connection establishment in inter-domain
networks. The service discovery using SLP protocol took around 7 ms and the CSPF
computation for the given topology with five NMS took around 0.252. In general the
method invocation such as SLA_Open(), SLA_Report(), and SLA_Decision() take
few milliseconds (30~40 ms).
2000
1782
1800
1600
1400
Time (ms)
1200
1000
800
600
400
200 7 38
0.252
0
Service CSPF Routing SLA Negotiation Association
Discovery - Method Traversals
Invocation
From the performance analysis, we found that the instance creation time depends
on the hardware specification and method invocation does not take much time. The
Provider API function for association traversals and enumerating instances also does
not take much time. We are developing multi-thread based parallelism in
implementation to improve the overall performance.
4 Conclusions
In this paper, we designed MOs for SLA negotiation for inter-AS TE-LSP connection
establishment with extensions on existing experimental DMTF CIM MOFs with
hierarchical inheritance. We designed MOs to represent SLA negotiations, service
registration/discovery, interASNet & QoSDiffServNet. We also designed MOs related
WBEM-Based SLA Management Across Multi-domain Networks 321
References
1. Young-Tak Kim, “Inter-AS Session & Connection Management for QoS-guaranteed
DiffServ Provisioning,” Proc. of International Conference on Software Engineering
Research, Management & Applications (SERA 2005), Mt. Pleasant, USA, pp. 325~330.
2. Web-based Enterprise Management (WBEM),
3. http://www.dmtf.org/standards/wbem/.
4. OpenSLP, www.openslp.org.
5. OpenPegasus, http://www.openpegasus.org/.
6. T.M.T. Nguyen, N. Boukhatem, G. Pujolle, "COPS-SLS Usage for Dynamic Policy-based
QoS Management over Heterogeneous IP Networks", IEEE Network, May/June 2003.
7. SLS Management, http://www.mescal.org/deliverables/d1.3_finalv2.pdf.
8. Goderis, D. et al., “Service Level Specification Semantics and Parameters”, Internet Draft,
<draft-tequila-sls-02.txt>, January 2002.
9. Dong-Jin Shin, Young-Tak Kim "Design and Implementation of Performance
Management for the DiffServ-aware-MPLS Network," Proceedings of Conference on
APNOMS 2003, Fukuoka, Japan, October 2003.
10. Shanmugham Sundaram, Abdurakhmon Abdurakhmanov, Young-Tak Kim, "WBEM-
based Inter-AS Traffic Engineering for QoS-guaranteed DiffServ Provisioning," IEEE
Broadband Convergence Networks (BcN2006) Workshop, Vancouver, British Columbia,
Canada, 2006.
Network Support for TCP Version Migration
1 Introduction
In recent years, speedup of backbone line progresses in the Internet, and, in
congestion controls of TCP (Transmission Control Protocol) used in the Internet
mainly, it is considered a technique to improve to be able to follow a broadband
of backbone line.
Currently a version called Reno is widely used for TCP about a congestion
control in the Internet. However, with a broadband of the Internet, it becomes
clear that we cannot deal with speedup of a network only by just using conges-
tion control algorithm of TCP Reno. Therefore there are number of researches
about new version of TCP to make improvement to TCP Reno. As such a new
TCP, HighSpeed TCP (HSTCP) [1]Fast TCP [2]Scalable TCP [3] are proposed.
HSTCP, for example, does increase spreading of a congestion window size greatly
for every one RTT in comparison with the one of Reno, and decrease small of
window size when a congestion occurred. It thus keeps a large quantity of win-
dow sizes during a communication. As a result, a HSTCP can support a line
speed of Gbps order.
However, because TCP is a congestion control protocol performed between
transmit and receive end nodes, it is necessary to replace the TCP protocol
stack running on the transmit or the receive (or sometimes both) node in order
to use a new version of TCP. As a result, more than one versions of TCPs coexist
in the same network line in the migration stage. At this time problem of fairness
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 322–331, 2006.
c Springer-Verlag Berlin Heidelberg 2006
Network Support for TCP Version Migration 323
about communication performance occurs between the users who use different
version of TCP, because difference in performance occurs between both versions
even under a similar network environment [4,5].
There are following two important issues to complete a step by step migration
between existing and new TCP versions.
1. A new version is to be good in performance compared to the existing version:
It is an end user or a service-provider (and not a network provider) that
judges whether it replaces TCP stack to a new version. There is no incentive
for users to shift a new version of TCP unless explicit effect (e.g., gain
additional throughput, or speed does not fall even if packet loss occurs) is
shown.
2. Not bringing remarkable performance degradation for a user using the ex-
isting version: A user (or service provider) cannot but become careful for
introduction of a new version from a viewpoint of stability use of network, if
a congestion control of a new version gives the one of the existing version big
effect, and, as a result, remarkable deterioration produces it in performance.
It has been considered much TCP versions by many researchers, but TCP
Reno is still used as the mainstream till now. It is because the effect that
would be occurred is not investigated enough in the environment where the
new version is in conjunction with TCP Reno. Currently there are enormous
nodes connected to the Internet, and it also indicates that the instability of
the control of TCP Reno directly means the instability of the Internet. From
the viewpoint of network operation, it is undesirable to do a version change
only by the reason of a respect to be efficient in performance.
These have a relationship to disagree each other. In other words, an incentive
to introduce a new version would be increasing according to its performance
advantage, while it would also be careful from the viewpoint of stability use of
a network. On the other hand, it can drive forward with safe when the com-
patibility between a new and the existing TCP versions is high, but if there is
no remarkable improvement in performance, merit of migration is not felt for a
user.
We propose a new model for TCP version migration by the network support
in this paper. We consider how a network provider (ISP) should promote a
migration to a new version of TCP with minimizing complaints from subscribers
who use the existing version. The advantage on a step-by-step migration of
TCP version by support of the network is that a new congestion control can be
designed more suitable without being conscious of an upper compatibility of the
existing version. It is possible that the network provider promotes the advantage
of a new version widely while minimizing complaints from existing version users.
We consider a step-by-step migration model of TCP version as shown in Fig. 1.
In this figure, the horizontal axis shows a time progress, and the vertical axis
shows the deployment share of each version. We can divide migration roughly
into following four phases.
324 S. Ata, K. Nagai, and I. Oka
1. Preliminary Stage: There are extremely a few users deploy a new version
experimentally. Because it is important that in this phase the advantage of
a new version should be recognized widely, we should not do any constraints
for a new version.
2. Migration Stage: The validity of a new version is recognized by some users,
and the number of users who updated to a new version is slowly. Fairness
with the existing version users becomes the problem in this phase. Therefore,
it becomes important to control complaints by the existing version users as
minimum as possible to recognize validity of a new version more widely.
3. Inversion Stage: When the merit of a new version is recognized widely, many
users are starting a change to a new version. It is expected that the number
of the updating users suddenly increases in this phase. It is because a new
version is implemented into standard function (APIs) of operating systems
(OS), and users can make a change to a new version by only invoking an
update of the operating system.
4. Final Stage: Most users have completed the change to a new version. We can
complete the change by recommending transference (in other words, update
of OS) to a new version for the existing version user individually.
Among the above-mentioned stage, it is migration stage that ISP has to support
mainly. It becomes important that we show an advantage in performance of a
new version while minimizing complaints of the existing version users. Therefore,
in this paper, we introduce subjective value (MOS:Mean Opinion Score) for a file
download as a metric to show the degree of complaints from the existing version
users. We then propose a new network support model to realize performance
advantages of a new version while achieving the MOS value where the existing
version users do not complaint. In this paper, we consider about step-by-step
migration from TCP Reno to HSTCP.
This paper is organized as follows. We first describe the impact HSTCP on
the migration stage in Section 2. We next model the relation between the perfor-
mance degradation for a file download and subjective scores in Section 3. Based
on these results, we propose a network support model to promote migration from
Network Support for TCP Version Migration 325
100 100
trial 1 trial1
90 trial 2 90 trial2
trial 3 trial3
Reno repletion rate (%)
80
100 100
Flow_num 10 Flow_num 10
90 Flow_num 11 90 Flow_num 11
Flow_num 12 Flow_num 12
Reno repletion rate (%)
80
Reno repletion rate (%)
80
Flow_num 13 Flow_num 13
70 Flow_num 14 70 Flow_num 14
60 Flow_num 15 60 Flow_num 15
Flow_num 16 Flow_num 16
50 Flow_num 17 50 Flow_num 17
40 Flow_num 18 40 Flow_num 18
30 Flow_num 19 30 Flow_num 19
Flow_num 20 Flow_num 20
20 20
10 10
0 0
0 10 20 30 40 50 60 70 80 90 0 10 20 30 40 50 60 70 80 90
HSTCP flow rate (%) HSTCP flow rate (%)
(a) RED (b) Drop-tail
5
MOS
Exponent function
4
MOS
2
0
100 90 80 70 60 50 40 30 20 10 0
Repletion rate(%)
By using both two relations (i.e., relation between the satisfaction ratio and the
ratio of HSTCP flows, and relation between the degradation (satisfaction) ratio
and the subjective score), we finally obtain the relation how the ratio of HSTCP
flows affects the MOS value. For simplicity we directly map the degradation
ratio to the satisfaction ratio. We also use the average value of satisfaction ratios
which are obtained by changing the total number of flows from 10 to 20, because
as previously described the satisfaction ratio is mainly affected by the ratio of
HSTCP flows (i.e., not by the number of HSTCP flows). By applying Eq. (1) we
derive the relation between the ratio of HSTCP flows and MOS values shown in
Fig. 7.
In this paper, we suppose that users feel complaint when the MOS value
becomes less than three. Under such definition we can observe that end users’
complaint occurs when the ratio of HSTCP flows becomes around 40% to 50%
of all flows in RED router case. However, the MOS value still remains about
2.4 even if the ratio of HSTCP flows becomes 90% of all flows. In other words,
RED routers are more suitable than Drop-tail routers when we consider the
migration of TCP versions. On the other hand, with Drop-tail routers, the MOS
value becomes less than 3 though the ratio of HSTCP flows is a few (i.e., 10%
to 20%). Moreover when the ratio of HSTCP flows reaches to 80% the MOS
value becomes less than 2, where most people have strong complaints on their
performance. From these results we propose a method to achieve the MOS value
for TCP Reno users more than 3 by the network support in the following section.
Network Support for TCP Version Migration 329
5 5
MOS MOS
4 4
3 3
MOS
MOS
2 2
1 1
0 0
0 10 20 30 40 50 60 70 80 90 0 10 20 30 40 50 60 70 80 90
HSTCP flow rate (%) HSTCP flow rate (%)
(a) RED (b) Drop-tail
5
Normal
Network Support
4
3
MOS
0
0 10 20 30 40 50 60 70 80 90
HSTCP flow rate (%)
forwarded packet with the random probability pd , i.e., the packet is forwarded to
the buffer of the router with the probability 1 − pd . The pass module is a single
packet buffer and simply passes the packet to the router buffer. The purpose
of the pass module is to keep the order of packet arrivals. After stored into the
buffer, the packet is forwarded to the next hop by the normal operation on the
router.
5 Concluding Remarks
In this paper, we have proposed a model for TCP version migration with the
network support. In this model we have introduced MOS which is the subjective
score with end users’ point of view. Through experiments we have clarified the
relation how the performance degradation make impacts to the subjective scores.
Based on the results we propose a model of network support with keeping the
target subjective score for TCP Reno users, and have shown that our model can
achieve the advantage of HSTCP with minimizing complaints from TCP Reno
users.
For future research topics, we need to improve the accuracy on the determi-
nation of packet dropping probability, and investigate the effect of the network
support in drop-tail router cases.
References
1. Floyd, S., Ratnasamy, S., Shenker, S.: High speed TCP for large congestion windows.
RFC 3649 (2003)
2. Wei, D., Jin, C., Low, S., Buhrmaster, G., Bunn, J., Choe, D., Cottrell, R., Doyle,
J., Feng, W., Martin, O., Newman, H., Paganini, F., Ravot, S., Singh, S.: Fast TCP:
From theory to experiments. IEEE Network (2005) 4–11
3. Kelly, T.: Scalable TCP: Improving performance in highspeed wide area networks.
In: Proceedings of ACM SIGCOMM. (2003) 83–91
4. Chuvpilo, G., Lee, J.W.: A simulation based comparison between XCP and high-
speed TCP. Laboratory for Computer Science Massachusetts Institute of Technology
(2002)
5. de Souza, E., Agarwal, D.: A highspeed TCP study:characteristics and deployment
issues. LBNL Technical Report LBNL-53215 (2003)
6. Rizzo, L.: Dummynet and forward error correction. Freenix 98 (1998)
7. Handa, Y., Minoda, Y., Tsukamoto, K., Komaki, S.: Measurements of utility for
latency time in wireless service and its dependence on users’ situation. IPSJ Journal
(in Japanese) 2005(47) (2005) 19–24
End-to-End QoS Monitoring Tool Development and
Performance Analysis for NGN
1 Introduction
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 332 – 341, 2006.
© Springer-Verlag Berlin Heidelberg 2006
End-to-End QoS Monitoring Tool Development and Performance Analysis for NGN 333
Fig. 1. Processor Architecture of the Terminal Agent for NGN QoS Measurement
The Quality Measurement Processor captures the packets that are required for quality
measurement (SIP and RTP / TRCP), analyzes the protocol header, and measures the
quality indicators (Connect Success Rate, One-Way Delay, Jitter, Packet Loss, R
value, MOS, and Resolution). It comprises the following blocks, with each block
defined as follows:
z SIP Session Initialization Block: This block creates a state management session
pool for the SIP session to get the quality indicators, such as the connect success
rate and resolution, organizes this into a linked list, and initializes the following
main session information required for the measurement of quality indicators:
{ SIP Key (Source IP/Port, Destination IP/Port), Source DN (Dial Number),
Destination DN, Call Phase (Fail, Success, In Process), Call Start Time, End
Time, SDP Media Information }
z RTP Status Initialization Block: This block creates a state table for the
RTP/TRCP packet process, organizes it into a linked list, and initializes the
following main RTP status table information required for the measurement of
quality indicators:
{ RTP Key (Source IP, Source Port, Destination IP, Destination Port), SSRC
Field Value, RTP Start/Final Sequence Number, RTP Start/Final Time Stamp,
336 C. Kim et al.
One-Way Delay (MIN, AVG, MAX) Value, Jitter Value, Total Received RTP
Packet Count }
z Packet Capture Block: This block captures the packet from the terminal, using
the packet capture library, and then sends it to the packet filter and classification
block.
z Packet Filter and Classification Block: This block filters and classifies the
packets required for quality measurement, and records the time stamp. After
checking the port number, it sends the SIP packet to the SIP processing block,
and the RTP packet to the RTP processing block.
z SIP Processing Block: This block receives the SIP packet from the packet filter
and classification block, analyzes the SIP header and SDP protocol, measures
the SIP session information (Source DN, Destination DN, Source IP, Destination
IP, Call Start Time, and Call End Time), call success rate, and resolution, and
records the result into the SIP session metrics of RTCP XR when the SIP session
is terminated. The call success rate (%) is measured by a formula, {(number of
total connected calls – number of failed-to-connect calls) / number of total
connected calls * 100}, determining as SUCCESS where the status code of the
SIP Response Message is 2xx or 3xx, and as FAIL where it is 4xx, 5xx, or 6xx
after receiving the SIP INVITE message and analyzing the procedure up until
200OK and the ACK message. Resolution, a measurement of the total number of
pixels displayed, is measured based on an SDP media attribute included in the
SIP response message.
z RTP Processing Block: This block receives the RTP/RTCP packet from the
packet filter and the classification block, analyzes the header, measures the
quality indicators (voice and video services end-to-end one way delay, jitter,
packet loss, R-value, and MOS value), saves them in the RTP state table, and
outputs the measurements of the RTP quality metrics of the RTCP XR message
at the SIP session is terminated. The quality indicator measurement method is
defined as follows: One-way delay, an end-to-end one-way delay to the packets
from the Send terminal to the Receive terminal, is measured by the formula
{RTT/2} after getting the RTT with the reference of the DLSR and LSR field
values of RTCP RR. Jitter, a variation of the end-to-end one-way delay to the
packets from the Send terminal to the Receive terminal, is measured according
to the formula, {J(i-1)+(Inter_arrival_jitter-J(i-1))/16}, where inter_arrival_jitter
= l(R(i)-R(i-1) -(S(i)-S(i-1))l, based on the RTP packet arrival time. The packet
loss rate is the rate of packet loss among the total packets of the actually
transmitted data, after a normal call connection has been made. The quality
indicator is measured by the formula, {(number of Send packets – number of
Receive packets) / number of Send packets * 100}, with the reference of the
RTP header sequence Number. R-value and the MOS value are measured using
a end-to-end objective, subjective quality evaluation method based on E-model.
For R-value is initially measured by applying the basic values of one-way delay,
packet loss, and codec type, which measured through an RTP header analysis, to
a formula suggested by the international standardization G.107. The MOS value
is measured based on a converted formula. For variables other than the three
items listed above, the basic values suggested by G.107 are used.
End-to-End QoS Monitoring Tool Development and Performance Analysis for NGN 337
Fig. 2 below illustrates a quality indicator measurement process after analyzing the
packets from each block of the quality measurement processor.
start
No
Yes No
RTP Packet Calculate Packet Loss Calculate Jitter RTP Voice
No Calculate R-Value
ICMP Packet
Yes Calculate MOS
The Measurement Result Processor receives the result from the Quality Measurement
Processor, sending it to the quality measurement server. It comprises the Transmit
Queue Initialization Block, the Measurement Result Receiving Block, and the
Measurement Result Transfer Block. Each block is defined as follows:
z Transmit Queue Initialization Block: This block creates a Message Send Buffer
to send the quality indicators from the QoS measurement processors to the QoS
measurement server.
z Measurement Result Receiving Block: This block creates an RTCP XR message
by reading the quality indicators stored in Transmit Queue by the Measurement
Result Transmit block of Quality Measurement Processor, and send it to the
Measurement Result Transfer Block. Fig.3. illustrates an RTCP XR message
format, which is extended in this paper, to send the quality indicators to the QoS
measurement server.
338 C. Kim et al.
z Measurement Result Transfer Block: This block sends RTCP XR quality metrics
from the Measurement Result Receiving Block to the quality measurement
server using the TCP/IP protocol.
0 1 2 3 4 5 6 7 0 1 2 3 4 5 6 7 0 1 2 3 4 5 6 7 0 1 2 3 4 5 6 7
header V=2 P Block Count PT=207 Length
Block Type = 100 Reserv Length
Source DN
Destination DN
Source IP
SIP Destination IP
QoS Call ID (128 Byte)
Metrics Call Start Time
Call End Time
Service Type Agent Type QoS Measurement Mode
OS Type CPU Speed
RAM Size Call Success Result
Block Type = 101 / 102 Reserv Length
SSRC
Total Sent Packets
Audio/
Total Receive Packets
Video
BPS (Kpbs)
QoS
One Way Delay (Min) One Way Delay (Avg)
Metrics
One Way Delay (Max) Jitter
R-Value MOS-Value
Resolution (Width) Resolution (Height)
This paper has suggested the quality measurement terminal agent for an NGN
environment, which uses the public packet capture library PCAP. For quality
measurement functions were implemented using Windows CE or Linux-based Ansi-C
language. In order to minimize the CPU load, specific modules were implemented
based on the Posix Thread. After being ported to a Linux or Windows-based
voice/video phone terminal, they are run as daemon type system services. Through
SIP/SDP protocol analysis, a variety of information (i.e.: session information, type of
service and terminal performance, call success rate, and resolution), can be measured.
The quality indicators, (i.e.: R-value, MOS value, delay, jitter, and loss), can also be
measured through an RTP/RTCP protocol analysis.
This paper has tested the quality measurement function to evaluate the performance of
the terminal agent for NGN QoS measurement. For the performance test, the terminal
agent was ported to a real-time Linux-based video phone and compared with the
common measuring instrument (QoS metrics) in terms of quality measurement
results. Fig. 4 below illustrates the testbed for performance evaluation of the terminal
agent for NGN QoS measurement:
End-to-End QoS Monitoring Tool Development and Performance Analysis for NGN 339
(a) one-way delay for audio (b) one-way delay for video
Fig. 5. Comparison of Measurement Result on One-way Delay
340 C. Kim et al.
(a) Loss for audio and video (b) R value for audio
Fig. 7. Comparison of Measurement Result on Packet Loss and R value
According to the performance test result above, it appears that the one-way delay was
less than 0.04ms on average, and that jitter was less than 0.05ms in terms of error
span. No error was observed, however, in terms of packet loss. R-value was found to
be less than 0.04 in terms of the error range. When the QoS measurement terminal
agent was implemented in a video phone terminal, roughly 400Kbyte was consumed
in terms of runtime memory use, while there was no influence determined on the
basic video telephone service. This result vindicates the notion that the NGN QoS
measurement terminal agent is reliable in terms of quality measurement performance,
and suitable for QoE-based end-to-end service quality management in an NGN
environment.
4 Conclusions
This paper has introduced the development of a QoS measurement terminal agent
through which QoE-based end-to-end service quality management is made possible,
and a performance test result is given. It has also demonstrated that the primary
quality indicators, such as R-value, MOS value, call success rate, delay, jitter, packet
loss, and resolution, may be measured by analyzing the packets on a regular basis
whenever the service is used with a video phone in which the QoS measurement
terminal agent is implemented. Based on the performance comparison with reliable
common tools, it has additionally been confirmed that the error rate is approximately
End-to-End QoS Monitoring Tool Development and Performance Analysis for NGN 341
the same. It has also been found that the agent is usable as a quality measurement tool
for QoE-based end-to-end quality management in an NGN environment.
Notwithstanding these findings, further research needs to be conducted to
compensate the function of the QoS measurement terminal agent through the
development of video MOS, and to make the function accessible in a small terminal,
such as a mobile receiver.
Acknowledgements. One of the authors of this work is supported by the second stage
BK21.
References
1. Handley, H.Schulzrine, E.Schooler, and J.Rosenberg, “SIP: session initiation protocol”,
RFC 2543, March 1999.
2. Wiley, "Internet Communications Using SIP Delivering VoIP and Multimedia Service
with Session Initiation Protocol”, 2001.
3. "TMF 701: Performance Reporting Concepts & Definitions ver.2.0", TM Forum, Nov.
2001.
4. "ITU-T G.107: The E-model, A computation Model for use in Transmission Planning",
2003.
5. Addison Wesley, "RTP: Audio and Video for the Internet", 2003.
6. ITU-T Rec. Y.1711, Operation & Maintenance mechanism for MPLS networks, Nov.
2002.
7. V. Paxson, G. Almes, J. Mahdavi, “Framework for IP Performance Metrics”, RFC 2330 ,
May, 1998.
8. R. Koodli, R. Ravikanth, “One-way Loss Pattern Sample Metrics”, RFC 3357, August,
2002.
9. C. Demichelis, P. Chimento, “IP Packet Delay Variation Metric for IP Performance
Metrics (IPPM)” RFC 3393, November, 2002.
10. V. Raisanen, G. Grotefeld, A. Morton, “Network Performance Measurement with periodic
streams”, RFC 3432, November, 2002.
11. EU Premium IP Project, http://www.cordis.lu/ist/ home.html
12. AQUILA Project, http://www.ist-aquila.org/
13. MESCAL, http://www.mescal.org/
14. TEQUILA, http://www.ist-tequila.org/
15. CAIDA, http://www.caida.org/
16. ChinChol Kim, SangChul Shin, Sunyong Han, “Architecture of end-to-end QoS for VoIP
Call Processing in the MPLS Network, 2004
”P4L”: A Four Layers P2P Model for
Optimizing Resources Discovery and
Localization
1 Introduction
Peer-to-Peer systems are distributed system without (or with a minimal ) cen-
tralized control or hierarchical organization, where each node is equivalent in
term of functionality. P2P refers to a class of systems and applications that
employ distributed resources to perform a critical function such as resources lo-
calization in a decentralized manner. The main challenge in P2P computing is
to design and implement a robust distributed system composed of distributed
and heterogeneous peer nodes, located in unrelated administrative domains. In
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 342–351, 2006.
c Springer-Verlag Berlin Heidelberg 2006
”P4L”: A Four Layers P2P Model 343
2 Related Work
Peer-to-Peer is relatively new in the areas of networking and distributed systems
and services (eg: Voice over IP )[9]. P2P computing started to be a hot topic
by the middle of years 2000. The different ”generations” of P2P systems are
characterized by transitions between generations motivated by different goals. In
this section, we describe and comment the different generations of P2P network.
– The first P2P generation like Napster file sharing application. The main
contribution of Napster was the introduction of a network architecture were
machines are not categorized as client and server but rather as machines that
offer and consumes resources. All participants have more or less the same
functionality. However, in order to locate files in a shared space, Napster
provides a central directory. Napster is composed of two services : a decen-
tralized storage service but with centralized directory service which can be
a single point of failure.
– The central coordination in the first solution leads to the transition to a
new kind of P2P system, where the focus is on the elimination of the central
344 M. Amad and A. Meddahi
coordination. This generation started with Gnutella application [4]. The sec-
ond generation systems solved the problem of the central coordination. How-
ever, the problem of scalability becomes more critical due to the network
traffic load generated by the flooding algorithm for research or localization.
Moreover, there is no guaranty to find a data item that exists in Gnutella
system, due to the limited search scope.
– In the third generation of P2P systems (initiated by research projects such
as Chord[10], CAN[6], Tapestry[12], Pastry[7]), P2P systems are based on
the Distributed Hash Table (DHT ) to generate keys for both nodes and
data. A node (Peer ) in such system requires a unique identifier based on a
cryptographic hash of some unique attribute such as its IP address. Nodes
identifiers and key value pairs are both hashed to one identifier space. The
nodes are then connected to each other in a certain predefined topology, eg: a
circular space (Chord ), d-dimensional cartesian space (CAN ). Our proposed
”P4L” model (A four levels P2P model ) is derived from Chord and used a
ring topology with hierarchical organization.
The common objective in all generation of P2P system is to optimize resources
discovery and localization in a dynamic and heterogeneous P2P system. The
next section describes this issue.
to a numeric key using an ordinary hash function such as SHA-1 or SHA-2. Then
a lookup (Key) function is called. Then the publisher sends the file or resource
to be stored at the resulting node (Data replication). The requester who wants
to access this file or resource, obtains the corresponding identifier, converts it to
a key, call the Lookup(Key ) function and ask the resulting node for retrieving
a copy of the file or resource.
”P4L” is a scalable P2P protocol aims to optimize resources discovery and
localization in a decentralized manner. It is based on DHT for resources iden-
tification and IP addresses for node identification. This model belongs to the
third generation of P2P systems which are based on specific topology. ”P4L”
is organized on four levels, each one is composed of several rings. One of the
most advantages of this proposed model is the rapidly convergence of the lookup
process. Section 5 describes, analysis and compares ”P4L” with others main P2P
protocols.
N90
72.15.124.240
N72
N69
K 81.92.10.23
N74
Level 1
N62
N81 N50
Level 2
K 51.20.30.1
N99 N10
81. 99. 2.16
N60
N41
N62
N123 N38
50.31.50. 125 50.31.50. 150
N209
50.31.50. 209
N90
finger table N50/N52
N50 + 1 N60
N72 N50 + 2 N60
Level 1
N74 N69 N50 + 4 N60
N50 + 8 N60
Level 1 N50 + 16 N69
N81 N62
N50
Level 2 Level 2
N52 N60
N99 N52 + 1 N60
N52 + 2 N60
Level 2
N60 N52 + 4 N60
N41 N52 + 8 N60
N52 + 16 N29
N50 N66
Level 2
N38
N29
N111
N120 N67
N31
Level 4
Level 3 Level 4
N61
N210
N200 N21 N52
N20
Algorithm 1 gives a simple illustration for the lookup process in ”P4L”. For
each level i ring, we use the ith part of the data key. If the request succeeds
on the ring k where the node requester belong, the cost of lookup algorithm
is O(ln(ni )), with ni the number of nodes in this ring (ni ≤ 256). In the case
where the request fails on the first ring, the search or localization is done on
ring level k + 1 or k − 1in a deterministic manner, then the cost of the lookup
process algorithm is O( (ln(ni ))), where ni is the number of nodes on ring level
i, where the request gets success.
topology can significantly and efficiently accelerate the lookup process. There are
Broadcast-based and DHT-based algorithms for P2P protocols[3], ”P4L” can be
considered as a combination of Broadcast-DHT-based P2P protocol. P2P net-
works tend to become a key element for Internet communications such as legacy
applications (eg: file sharing) but also for VoIP [9]. However, security aspects
constitute a serious concern for P2P. (eg. in ”P4L” context, a malicious node
can present an incorrect view of ”P4L” lookup process). For taking into con-
sideration security aspect in ”P4L”, we are interested to combine and extend
security protocols in the context of ”P4L” for large peer group communications.
References
1. Curt Cramer, Kendy Kutzner, and Thomas Fuhrmann. Bootstrapping Locality-
Aware P2P Networks. 2003.
2. C.Shirky. What is P2P..and what Isn’t. O’ Reilly Network, 2001.
3. Gang Ding and Bharat Bhargava. Peer-to-Pee file sharing over mobile Ad hoc
Networks. Proceedings of the second IEEE annual conference pervasive computing
and communications Workshops (PERCOMW’04), 2004.
4. Gnutella. http://www.gnutella.com.
5. Dejan S. Milojicic, Vana Kalogeraki, Rajan Lukose, Kiran Nagaraja, Jim Pruyne,
Bruno Richard, Sami Rollins, and Zhichen Xu. Peer to Peer conputing Survey. HP
Laboratories Palo Alto, HPL-2002-57, March 2002.
6. Sylvia Ratnasamy, Paul francis, Mark Handley, Richard Karp, and Scott Shenker.
A scalable content addressable network. ACM SIGCOMM, 2001.
7. Antony Rowstron and Peter Druschel. Pastry : a scalable, decentralized object
location and routing for large scale peer to peer systems. Proceedings of the 18 th
IFIP/ACM international conference on distributed systems plateforms (Middelware
2001), Heidelberg, Germany, November 2001.
8. Detlef Schoder and kai Fischbach. The Peer-to-Peer Paradigm. Proceeding of the 38
th Hawaii International Conference on System Sciences, IEEE Internet computing,
2005.
9. Kundan Singh and Henning Schulzrinne. P2P Internet telephony using SIP. Tech-
nical report, Department of computer Science, Columbia University, 2004.
10. Ion Stoica, Robert Morris, David Karger, M.Frans Kaashoek, and Haris Balakr-
ishnan. Chord: A Scalable Peer-to-Peer lookup Service for Internet Application.
Proceedings of the ACM SIGCOMM’01, san Diego, California, 2001.
11. Peer to Peer Working Group. Bidirectional Peer-to-Peer communication with in-
terposing Firewalls and NATs . Peer-to-Peer Working group, White Paper, 2001.
12. Ben.y. Zhao, John kubiatowich, and Anthony D. joseph. Tapestry: an infrastructure
for fault-tolerant Wide-area location and routing . Report No. UCB/CDS-01-1141,
Computer Science Division, University of California, Berkeley, April 2001.
A Zeroconf Approach to Secure and Easy-to-Use Remote
Access to Networked Appliances
1
KDDI R&D Laboratories Inc., 2-1-15 Ohara Fujimino-shi Saitama 356-8502, Japan
2
KDDI Corporation, 3-10-10 Iidabashi Chiyoda-ku Tokyo 102-8460, Japan
1 Introduction
As we see always-on broadband access such as Fiber To The Home (FTTH), wide
spread of mobile phones and emerging smart networked appliances (NAs), a variety of
technologies have formed the basis of home networking. We can now have a
connection anywhere from home and vice versa, as well as at traditionally limited
universities or research institutes. This will steadily evolve into our new environment
known as ubiquitous networking, which will support us by accessing everything
connected together anytime and anywhere.
In such an environment, NAs would be a key driver for one of new multi-play
services: remote access service. Typical NAs are a network camera, network attached
storage (NAS) and network DVD player, which are abbreviated NW camera, NAS and
NW DVD player in brief hereafter. With these NAs reachable from the Internet, we can
check if a key to the front door is closed from a place away, by access to our NW
camera facing the door, using mobile phone. Another scenario is remote recording
reservation. We can make a recording reservation of a TV program, by access to our
NW DVD player, using a mobile phone.
Some architecture and their associated protocols for NA remote access services have
been developed and standardized. Typical protocols are based on Session Initiation
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 352 – 361, 2006.
© Springer-Verlag Berlin Heidelberg 2006
A Zeroconf Approach to Secure and Easy-to-Use Remote Access 353
Protocol (SIP). Unfortunately, they still stay at specifications and lab-scale prototypes
and neither could they fully address the following technical issues: (1) configuration
task, (2) usability nor (3) deployability, which will be described more in Sect.3.
For a solution to the issues, this paper proposes a new approach to secure and
easy-to-use remote access to NAs. Servers in an Internet Service Provider (ISP) domain
and a residential gateway (RGW) in a home network play an important role. The secure
access is ensured by Security Architecture for Internet Protocol (IPsec) tunnel between
an ISP domain and a home network. The easy-to-use access, which is one of our main
contributions, is based on two zeroconf protocols: (1) Universal Plug and Play (UPnP)
and (2) a new registration protocol. The approach has two stages: discovery and
registration. In the discovery stage, a RGW discovers a new NA connected to a home
network and monitors its connectivity using Simple Service Discovery Protocol
(SSDP) and Simple Object Access Protocol (SOAP). In the registration stage, a RGW
registers and updates a UPnP device description of connected NAs with servers using
the new registration protocol. In addition, the zeroconf protocols are so designed that
they can ensure the interoperability with UPnP noncompliant IP-based NAs. With the
servers working as a rendezvous point, the proposed approach allows even Internet
novices to access typical NAs available in the marketplace, using mobile phone and PC
anytime and anywhere, with minimum user intervention.
An emphasis of the paper also lies in development and deployment of a system
based on the proposed approach to show its practicality. We implement the zeroconf
protocols in the form of RGW firmware and install the servers to work with the RGW,
taking the current typical home network into account. For the practicality, we evaluate
the proposed approach in terms of the two zeroconf protocols empirically through a
field trial, in which approximately 1700 users took part in for four months, conducted
as part of the deployment.
This paper is organized as follows: In Sect.2, we present an overview of a typical
home network with FTTH connections and show technical issues for the NA remote
access. We review the recent related work in Sect.3. In Sect.4, we propose a new
approach to secure and easy-to-use remote access to NAs. In Sect.5, we describe
implementation of the system, and evaluate the proposed approach empirically through
the field trial.
Figure 1 shows a typical home network, which we focus on throughout the paper, with
fiber-optic network operator and ISP domains. Consumer premises equipment (CPE)
devices including a media converter, RGW, PCs, and phone are connected in a tree to
the RGW as its root. NAs are connected to the RGW in the same way. The RGW works
as an IP router and has a Dynamic Host Configuration Protocol (DHCP) server for the
home network. It has also Voice over IP (VoIP) capability. A home network is
connected to the ISP domain, in which World Wide Web (WEB), Simple Mail Transfer
Protocol (SMTP), Post Office Protocol (POP), SIP and Video on Demand (VoD)
354 K. Yoshihara, T. Maruta, and H. Horiuchi
servers are operated to serve such as triple play services via the fiber-optic network
operator domain.
The following technical issues for the NA remote access should be addressed.
1. Reachability from the Internet to NAs should be ensured (Issue 1).
2. Communication with NAs should be authenticated and encrypted, to protect a
network and privacy from unauthorized access by malicious or erroneous operation, for
NAs are generally tied with individual activities (Issue 2).
3. An NA should be configured and discovered automatically, with minimum user
intervention, to release home users from the complex configuration task and to inform
users of the presence of dynamically connected NAs (Issue 3).
4. More usability by which an NA is given an associable name like ``Network
Camera'' instead of a usual form of ``http://192.168.0.1/xxx.cgi'' should be achieved, so
that users can select a connected NA with ease (Issue 4).
5. Architecture and protocols to realize the remote access should be subject to the
current typical home network and to interoperability with NAs available in the
marketplace, for the rapid deployment and acceptance (Issue 5).
3 Related Work
Research and development work on the NA remote access have been found. We
summarize recent related work below and show that none of them alone addresses all
issues in Sect.2 simultaneously.
The work propounded the use of SIP for the secure protocol of the NA remote access
and discussed requirements and future challenges. Unfortunately, the work stayed at
prospects and did not show how to address the requirements and challenges except
Issue 1 and 2 in depth technically, while their contributions might be a few architectural
frameworks and SIP message extensions.
Ubiquitous Open Platform Forum (UOPF) founded by Japanese companies has
developed and standardized the UOPF protocol with its base architecture. The protocol
A Zeroconf Approach to Secure and Easy-to-Use Remote Access 355
addresses Issue 1, since it is also based on SIP and a user can access a target NA from
the Internet using SIP as a signaling protocol. Network Address Translation (NAT)
may be traversed with the aid of UPnP. The protocol meets Issue 2, for some extensions
have been made to ensure the peer-to-peer secure communication using IPsec and
Transport Layer Security.
Both protocols, however, could address none of the other issues due to their reliance
on SIP. With respect to Issue 3, for each NA, a user should initially configure SIP
Uniform Resource Identifiers (URIs) typically in the form of ``sip:[email protected]'' to
identify itself and gain access to SIP servers. As for Issue 4, due to the URIs typically
given by an ISP as well as an email account, an NA cannot necessarily have an
associable name, leading to poor usability. In terms of Issue 5, the protocol imposes SIP
support on every NA. It is hard to find such an NA in the marketplace except for SIP
phones as of writing, although the UOPF protocol provides specifications of a gateway
for noncompliant NAs.
4 Proposed Approach
Design Principles. The proposed approach is designed based on the four principles as
shown in Fig.2.
terminal selects a target NA on the pages. The request is finally routed to the NA
through the tunnel. Source of the pages is provided by the new protocols below.
2. New protocols for auto-discovery and auto-registration
To tackle Issue 3, the RGW leverages DHCP, IPv6 Stateless Address
Autoconfiguration (SAA) and SSDP M-SEARCH to auto-configure and
auto-discover NAs that may be dynamically connected to a home network. The
RGW obtains UPnP device descriptions from NAs, using SOAP after the discovery.
In addition, a new auto-registration protocol based on Hyper Text Transfer Protocol
(HTTP) by which the RGW registers the device descriptions with the servers is
introduced to provide the web pages as described above. The detail of the protocol
will be shown in Sect.4.2.
3. UPnP device description for more usability
To address Issue 4, the web pages for the NA remote access is provided such that an
associable string like ``NetworkCamera'' specified with the ``friendlyName'' tag in
the UPnP device description is used as a link string. A URL specified with the
``presentationURL'' tag, which corresponds to a usual ``index.html'' of an NA, is
used as a reference of the link. The string on a terminal may be grayed out and
cleared, when a corresponding NA is disconnected.
4. RGW coping with UPnP noncompliant IP-based NA
The RGW exploits Internet Control Message Protocol (ICMP) and ICMPv6 Echo
requests to discover and register UPnP noncompliant IP-based NAs. The portal
server allows a user to configure an associable name and TCP/UDP ports for those
NAs.
With the four principles, we address Issue 5 in that 1) a RGW widely deployed in
home networks and always operating for VoIP incoming call serves the discovery and
registration, and 2) interoperability with both UPnP compliant and noncompliant NAs,
available in the marketplace is achieved.
Assumptions. We put the following assumptions for the proposed approach.
1. The configuration of a RGW has been completed properly. Reachability from a
home network to a subscribing ISP domain is ensured.
2. Servers for the NA remote access are installed. Moreover, a server that maintains
versions of RGW firmware is set up in the same ISP domain. A RGW visits the
server and checks if the current version is the latest, regularly or when the RGW
(re)boots, so that it can keep its firmware up-to-date.
3. A user account and its associated password necessary to establish an IPsec tunnel
are enrolled in the user DB by operators in advance.
4. A new version of RGW firmware for the proposed approach is released. It can
establish an IPsec tunnel from a RGW to the VPN server automatically, and can
then perform the auto-discovery and auto-registration. All required settings: IP
addresses of the servers, a user account and password for the tunnel are preset,
using a subscriber account as a key.
A Zeroconf Approach to Secure and Easy-to-Use Remote Access 357
5. A RGW updates the current firmware with the above new version.
6. A user creates a user account and its associated password for the remote access,
when the user first visits the portal server from a home network.
Components in ISP Domain. Four server components: a portal, DNS, user DB and
VPN servers constitute an ISP domain as shown in the left side of Fig.3.
The portal server operates as an HTTP server. It manages and provides web pages
linking to NAs connected to a home network for each user, and processes the NA
remote access requests. It generates a pair of a Fully Qualified Domain Name (FQDN)
and its associated IPv4 or IPv6 address for each NA, on receiving a registration request
from a RGW, to resolve duplication caused by the same IPv4 address space among
users. The server allows a user to customize preferences: to create, delete and change a
user account and password for the remote access, to control access to NAs based on the
user account and to configure an associable name and TCP/UDP ports for UPnP
noncompliant IP-based NAs. With the DNS DB, the DNS server is responsible for the
registration of pairs of an FQDN and its associated IP address, when it receives a
request from the portal server, as well as the original name-to-address resolution. The
user DB server manages user related information: a subscriber account, pairs of a user
account and password to establish an IPsec tunnel and for the remote access, and part of
the source of the web pages. The VPN server establishes and releases IPsec tunnels
based on requests from a VPN client in a home network.
specified in device descriptions, in the form of a logical NA table. The RGW monitors
NA connectivity by regular M-SEARCH request, and updates the NA table. The RGW
deletes an entry of the NA table if it does not receive a specified number of consecutive
responses from a corresponding NA, leading to clearance from a screen via gray out. In
terms of ICMP, addresses to which a RGW sends the requests can be limited with a
start address and a number of addresses, such as ``192.168.0.2'' and ``6'', to avoid
bandwidth consumption. With respect to ICMPv6, a RGW first sends a request to the
link local multicast address ``FF02::1''. When the RGW receives a reply, it sends a
subsequent request to a global address that is generated based on SAA. On receiving an
ICMP or ICMPv6 Echo reply, the RGW stores a MAC and IP address of the NA, and
monitors in the same way as M-SEARCH request. Two or more entries may be stored
for an NA, if the NA operates on IPv4/v6 dual stack and replies to both ICMP and
ICMPv6 Echo requests. Such duplication will be resolved on the portal server, using
MAC address as a key.
Auto-registration function registers the NA table with the portal server on a regular
basis, using a newly introduced protocol based on HTTP. An example of the protocol
payload is shown in Fig.4. The RGW part includes Wide Area Network (WAN)
addresses of the RGW. The NA part is defined as a list. Each element corresponds to an
entry of the NA table. The list can contain an arbitrary number and type of elements,
being independent of UPnP compliance.
The VPN client requests the VPN server to establish and release an IPsec tunnels. It
sends keep-alive messages for the tunnel maintenance.
We describe the two flows of the proposed approach with Fig.3, to show how it allows
for the secure and easy-to-use remote access.
registration request generates pairs of an FQDN and an IP address for NAs, and
registers the pairs with the DNS server (Fig.3(6)). After successful registration
(Fig.3(7)), the portal server stores the up-to-date information including the payload as
shown in Fig.4 in the user DB server (Fig.3(8)).
Remote Access Flow. When the portal server receives a remote access request from a
user (Fig.3(a)), it authenticates the user with the aid of the user DB server (Fig.3(b)).
Success of authentication (Fig.3(c)) provides the user with web pages for the remote
access via the portal server (Fig.3(d)). When the user selects a target NA (Fig.3(e)), the
portal server creates a query for name-to-address resolution to the DNS server
(Fig.3(f)). The request is routed to the resolved address (Fig.3(g)) through the IPsec
tunnel (Fig.3(h)).
5.1 Implementation
We implement a system based on the proposed approach and describe a brief overview
of the system below, including some restrictions on the implementation.
1. We implement the servers on separate hardware, using general-purpose server.
The servers are connected with 100Base-TX Ethernet in adjacent LANs, while
we implement the DBs, using a single storage (4.5Tbyte in physical) that is
connected to its associated server with fiber-channel. The specification of the
portal server in terms of a quadruplet (CPU(GHz), memory(Gbyte),
HDD(Gbyte), OS) is (UltraSPARC IIIi 1.28 2, 2, 73.4 2, Solaris9.9).
2. We implement a new version of RGW firmware for the proposed approach. The
target RGW is models that have already operated in home networks.
3. We implement the VPN client software that runs on a user PC, due to some
models that are restricted in processing capability. In this case, to establish an
IPsec tunnel, a user should enter an account and password on a PC.
4. Interoperability with NAs is ensured, such as NW camera, NAS and NW DVD
player, shipped from multi-vendors and available in the marketplace.
We deployed the system in Sect.5.1 with which a commercial FTTH network for a triple
play service was overlaid, to conduct a field trial for four months. We provided the
firmware to approximately 1700 users who were the FTTH subscribers and applied for the
trial. We recommended the use of NAs that were ensured the interoperability to the users.
The amount of messages for the auto-discovery by the system in 5.1 will be shown
first, in order to evaluate the traffic overhead. Part of server activities in logs during the
trial will be shown next, so as to evaluate the overhead of the auto-registration and to
estimate the scalability towards commercial deployment. Some discussions to improve
the system will be made after each evaluation.
360 K. Yoshihara, T. Maruta, and H. Horiuchi
400
730 2902
NAS (UPnP Compliant and IPv4 Support) 300
744 2812
NW DVD player (UPnP Compliant and IPv4 Support) 200
954 2267
NW Camera (UPnP Compliant and IPv4/IPv6 Support)
100
From RGW to NA From NA to RGW
Evaluation of Auto-Discovery. Figure 5 (a) shows the amount of messages for the
auto-discovery, when we connect each NW camera, NAS and NW DVD player to the
RGW with 100Base-TX Ethernet. Due to a device description, the amount of messages to
RGW is more than that from RGW for every NA. The amount of messages to the NW
camera is more than those of the other NAs, for an ICMPv6 Echo request and reply. The
total management traffic is 83.5bps ( = (730+2902+744+2812+954+2267)*8/60) when
the monitoring interval is 60 seconds, which was the same as in the trial. This is
negligibly small, for it is much less than bandwidth of a typical home network: 100Mbps.
We can shorten the interval to provide users with a more consistent view of dynamically
connected NAs, since the response time from an M-SEARCH request is at most
421.6msec. In addition to such regular discovery, we will be able to utilize ``ssdp:alive''
and ``ssdp:byebye'' message, although listening to such message required more
processing capability from the RGW than that at the design phase.
Evaluation of Auto-Registration. Figure 5 (b) shows the inbound traffic of the portal
server monitored every five minutes for the first month. It includes only the traffic used
for the remote access. At least around 80Kbps traffic arriving constantly from the third
week can be seen for the auto-registration, where its time interval was 10 minutes. The
amount of messages for the auto-registration per user is estimated at approximately
14.1Kbyte (=80*10*60*/(1700/4)/8), if we can assume one fourth of users had already
joined. This is acceptable, when we recall an ISP-style email service. A huge number of
users are sending pile of emails of the size that is typically larger than that, due to rich
contents attached.
An unavailable time at the end of the third week was caused by a planned network
construction. Redundant configuration in the ISP domain in terms of both hardware and
software will be required toward the scalable commercial deployment, to avoid such a
single point of failure.
6 Conclusions
This paper proposed a new approach to secure and easy-to-use remote access to
networked appliances (NAs). We show five technical issues for the NA remote access.
Lack of a protocol addressing all issues simultaneously motivated us to design a new
A Zeroconf Approach to Secure and Easy-to-Use Remote Access 361
Acknowledgment
We are indebted to Dr. Shigeyuki Akiba, President, Chief Executive Officer of KDDI
R\&D Laboratories Inc., for his continuous encouragement to this research.
References
1. Moyer, S., Marples, D., Tsang, S.: A Protocol for Wide-Area Secure Networked Appliance
Communication. IEEE Communications Magazine (2001) pp.52–59
2. Moyer, S., Marples, D., Tsang, S., Ghosh, A.: Service Portability of Networked Appliances.
IEEE Communications Magazine (2002) pp.116–121
3. Ubiquitous Open Platform Forum (UOPF): UOPF Protocol. (2004) http://uopf.org/en/
(URL available on April 2006).
4. Rosenberg, J., et al.: SIP: Session Initiation Protocol. IETF RFC 3261. (2002) 5. UPnP
Forum: Universal Plug and PlayTM Device Architecture. (2000)
5. World Wide Web Consortium: SOAP Version 1.2. (2003)
6. Thomson, S., Narten, T.: IPv6 Stateless Address Autoconfiguration. IETF, RFC 2462.
(1998)
7. Conta, A., Deering, S., Gupta, M.: Internet Control Message Protocol (ICMPv6) for the
Internet Protocol Version 6 (IPv6) Specification. IETF, RFC 4443. (2006)
A Test Method for Base Before Service (BS) of
Customer Problems for the NeOSS System
Abstract. As a network & service provider, KT has long felt the necessity of
integrating many of its OSS (Operation Support System) systems, and in 2003,
KT finally developed a new system to serve its various maintenance activities.
This system is called the NeOSS (New Operations Support System) System,
which uses XML (Extensible Markup Language), EAI (Enterprise Application
Integration) and Web service based on Microsoft’s .NET platform technology.
This paper shows the system architecture of the NeOSS-ADM(Access Domain
Management), which is going to be developed using the .NET environment,
service maintenance framework and the method of maintenance for NEs
(network elements) for base BS. And this system will give more high quality
service to customers and reduce the customer's complaints.
1 Introduction
Korea is the country with the highest distribution rate for high-speed internet service
in the world, and in recent years, its customer base is becoming saturated. The number
of high-speed internet customers in Korea has grown 12.26 million (05.12). KT alone
accounts for more than 6 million customers. Thus, KT is confronted by keen
competition due to the saturation of customers and the emergence of diverse service
providers that offer high-speed and broadband network services. In addition, customer
requests have become more diverse, from service orders to complaints related to fault
resolutions, quality of service, cost, rapid processing, and others.
In recent years, telecommunication services are being provided through complex
networks instead of single networks, particularly the various internet services. Thus,
many companies have developed many kinds of new services in the desire to obtain
larger market shares. To give the general public a high-speed internet access service,
KT supports various access technologies, including PSTN (Public Switched
Telephone Network), ADSL and VDSL, in order to provide broadband services
through copper wire. KT had also established an evolution plan for optical fiber
access networks. In accordance with this plan, KT developed and deployed the FLC
(Fiber Loop Carrier) –A and -B systems for business applications (FTTO) and the
FLC-Curb(C) and –Dense(D) systems for densely populated areas (i.e. apartment
complexes) and small business applications (FTTC).
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 362 – 371, 2006.
© Springer-Verlag Berlin Heidelberg 2006
A Test Method for Base Before Service (BS) of Customer Problems 363
KT also recently developed an AGW (Access Gateway) for the BcN (Broadband
Convergence Network). In the near future, KT will also deploy a soft-switch for the
BcN. The FLC-C, -D and AGW systems are able to connect to network switching
systems that support the ETSI V5.2 standard.
KT also needed a new paradigm for its OSS system because of the continuous
emergence of new products (Internet services) and extensive customer needs. In
particular, customers’ satisfaction level for service quality is much higher than before.
Thus, KT concluded that these issues cannot be resolved without basically changing
its operational management system. Therefore, KT developed the NeOSS system to
establish a new, professional and flexible architecture system. The NeOSS system is
an integration of the telecommunication management system and rebuilding system
per facility part with the management process TMN (business management, service
management, network management, equipment management) based on other coupled
systems. The NeOSS manages all of the company’s information relating to quality,
facility, traffic, customer service, and others.
2 NeOSS Structure
Our system’s goal is to support a one-stop fault management scheme and building
system. We can check where the NeOSS-ADM is on the NGOSS eTOM model and
how we can develop our system to co-work with any other system or facility.
Fig. 1. The NeOSS Architecture is harmonized with the NGOSS Framework and eTOM
The NGOSS is driven by TMForum for the standard business solution framework
of the next operation support system (OSS)/business support system (BSS). The main
contents of the NGOSS are business process modeling, design process, choice
technology, public data modeling and implementation. eTOM is the general business
364 I. Hwang, S. Seok, and J. Yoo
activity map that service providers must provide and the business process model
enterprise business process framework that service providers must request. The
eTOM model decides the borderline of software systems and components supplied by
the OSS/BSS. A concluding remark is that the eTOM model provides liaison with
future NGOSS and improves the relationships of businesses under the more complex
circumstance of e-business.
We applied a CBD (component-based development)-based architecture to
minimize the dependency between components and also increased the reusability to
reduce development costs. The NeOSS-ADM locates the SML layer and “Service
Assurance” part in the eTOM business process model.
into a few and simple categories and chose them to achieve inter-operability with
legacy systems.
Most Service Providers focus on one or two offerings when introducing new
services. As a result, similar functions are reproduced by various OSS/BSS
applications. This kind of limited scope results in offerings that are very poorly
coordinated and have few synergies. They also take longer to build new services
(almost starting from beginning each time) and new services are expensive to
introduce. To solve this problem, we defined the principle and mechanism of inter-
operation with other OSS/BSS. The inter-operation mechanisms are as follows:
Web Service (Standard)
Web Service (Protocol Adapter)
EAI (Standard)
EAI (Protocol Adapter)
Protocol Adapter mechanisms were applied for inter-operating with legacy systems
such as various NMS or BSSs. In this mechanism, we provide adaptors such as the
socket-based method, database-based method, XML-based method and others.
Standard mechanisms were applied for inter-operating between the NeOSS
subsystems, such as NeOSS-SO, NeOSS-SA, NeOSS-ADM, NeOSS-WM and
NeOSS-FM.
In Figure 4, we show part of the web service (protocol adapter) for NeOSS’s inter-
operation with other systems. The web service method consists of adapters, such as
socket, XML-RPC, HTTP and Oracle Database adapters. Through a sync or Async
way, the web services also decide on whether or not to use the EAI bus (the EAI bus
basically uses the Async way). Figure 4 shows its inter-operation with other systems.
KT started to develop the NeOSS (New Operation Support System) to reinstall the
communications net and network management processing module and provide a new
and advanced flexible architecture system in 2003. The NeOSS system underwent
A Test Method for Base Before Service (BS) of Customer Problems 367
The EMS Monitoring function monitors the real-time state of the EMS for each
network element. We send the specific message, which are hello and ping, to the
EMS. The ping message checks the network status from the NeOSS-ADM-TM to the
EMS. If we get a result of fail, we conclude that the network is not connected from
the NeOSS-ADM-TM to the EMS. Otherwise, we conclude that there is no problem
with the network status.
368 I. Hwang, S. Seok, and J. Yoo
The hello message checks the specific processing module of the EMS, which
connects to the NeOSS-ADM-TM system’s processor. If we get a result of fail, we
conclude that the EMS’ process is dead. So, we call the manager who manages the
EMS system. In addition, we can set the period test for Hello & Ping. For example,
we can select a period of 5 or 30 seconds, and then the NeOSS-ADM-TM system
sends the Helloand Ping messages to the EMS every 5 or 30 seconds.
There is a lot of information available from the GUI screen. It displays the
manager’s telephone number and mobile telephone number, the result of the status,
which are office, IP, type of NE, maker, status of EMS and test time of each EMS
test, and the history of the status of the EMS test. There is also a special function for
sending an SMS (short message system) to the manager.
This displays the history of the instant test and the detailed test, the information
displayed screen order by office, service, type of NE and maker. Furthermore, it can
perform an instant test for the specific customer’s number.
The operator wants to know the history of each NE and the method for solving the
problem in order to prevent the same mistake in managing the EMS or to analyze the
status result of the EMS. And he wants to level up the test success rate using this
function.
METF also manages the test result data of each EMS sorted by headquarters,
central office, office, service, EMS and maker. One is the test rate per office, which
displays statistics on the office’s test success rate and the chart to understand the test
rate transition of the office. The other is the test rate per EMS, which displays
statistics about each EMS’s test rate and the chart to understand the test rate transition
of each EMS. And the operator can ask for the result of each testing history and save
an Excel file when he wants to save this content to a file.
A Test Method for Base Before Service (BS) of Customer Problems 369
5 Proactive Maintenance
the stage of trouble processing, the operator executes the detailed test and dispatches
the technician.
5 Conclusion
The major features and functions of the NeOSS-ADM-TM system, which provides
NE maintenance for the maintenance of the normal status of NE’s EMS and proactive
maintenance for repairing customer faults, are described at this paper.
The NeOSS-ADM-TM performs the test for each NE and subscriber line using the
TU (Testing Unit) or inter-operation with the NE’s EMS. Then, the operator requests
for the test result and they compare the normal condition. If the operator concludes
that the test result is fault, the operator turns over this customer to trouble processing.
In the stage of trouble processing, the operator executes the detailed test and
dispatches the technician.
Nowadays, customer requirements for high-level quality PSTN or ADSL services
are more varied, and there are a lot of new and varied telecommunication network
devices. Many other services and new facilities will also appear in the near future, so
we have to adapt to that. Therefore, KT has designed the NeOSS-ADM, which
integrates the access domain system of service provision, service management and
network management.
The developed NeOSS-ADM-TM improves customers’ trust in our company, KT,
thus giving us the competitive power over any other company’s service quality. The
NeOSS-ADM-TM also supplies the basis for Before Service for all customers.
There is a need to further study the security issues for the NeOSS-ADM-TM and
promote the test success rate & test correctness rate. We are preparing new algorithms
for test correctness rate and will deploy a new version of the NeOSS-ADM-TM later
this year.
A Test Method for Base Before Service (BS) of Customer Problems 371
References
1. Shin-ho Choi: Building a Service Assurance System in KT, NOMS, (2004)
2. Bok-Gyu Hwang: Integrated Network Management System for Access Network in KT,
APNOMS, June (2003)
3. Mae-Hwa Park: The NeOSS-ADM development, Korea Telecom Technical Review,
Vol.17, No 4, Dec (2003)
4. TMForum: New Generation Operations Systems and Software (NGOSS) Architecture,
March (2003)
5. TMForum: Enhanced Telecom Operations Map (eTOM®): The Business Process
Framework-for the Information and Communications Services Industry, June (2002)
6. Hyunmin Lim: Web-based Operation Support System for the maintenance of Access
Networks in Korea Telecom, NOC2000, June (2000)
Self-management System Based on Self-healing
Mechanism*
1 Introduction
Self-management entails many different aspects, resulting in distinct dimensions of
control. For instance, IBM’s autonomic computing initiative views self-managing
systems as typically exhibiting a subset of four capabilities: self-configuring (adapt
automatically to dynamically changing environments), self-healing (discover,
diagnose, and react to disruptions), self-optimizing (monitor and tune resources
automatically), and self-protecting (anticipate, detect, identify, and protect themselves
from any attacks) [1].
As the complexity of distributed computing systems increases, systems
management tasks require significantly higher levels of automation. Thus, there is a
growing need for experts who can assure the efficient management of various
computer systems. However, management operations involving human intervention
have clear limits in terms of their cost effectiveness and the availability of human
resources [1]. Of all computer problems, about 40% are attributable to errors made by
the system administrators [2]. Thus, the current system management method, which
depends mainly on professional managers, needs to be improved.
To solve these problems when they do arise, there is a definite need for effective
self-healing system. The existing self-healing systems consist of a 5-step process,
*
This work was supported in parts by Ubiquitous Autonomic Computing and Network
Project, 21th Century Frontier R&D Program, MIC, Korea under ITRC IITA-2005-(C1090-
0501-0019), and grant No, R01-2006-000-10954-0 from the Basic Research Program of the
Korea Science & Engineering Foundation.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 372 – 382, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Self-management System Based on Self-healing Mechanism 373
2 Related Works
Oreizy et. al. [4] proposed the following processes for self-adaptive software:
Monitoring the system, Planning the changes, Deploying the change descriptions and
Enacting the changes [5][6][7][8]. The Adaptive Service Framework (ASF) [10]
proposed by IBM and CISCO consists of a 5-step process, including Monitoring,
Translation, Analysis, Diagnosis and Feedback. These 5 processes are applied in the
form of self-adaptive behaviors. The functions of the ASF are as follows: firstly, the
Adapters monitor the logs from the various components (Monitoring). Secondly, the
Adapter [9] translates the log generated by the component into the CBE (Common
Based Event) format (Translation). Thirdly, the Autonomic Manager [9] analyzes the
CBE log. This step identifies the relationship between the components through their
dependency (Analysis). Fourthly, the Autonomic Manager [9] finds the appropriate
healing method by means of the Symptom Rule [9] and Policy Engine [9] and then
applies the healing method to the applicable component. The feedback from the
Resource Manager [9] enables the system to heal itself (Diagnosis and Feedback).
Finally, in the event that the component has a critical problem or one which cannot be
solved easily, the Autonomic Manager sends a Call Home Format1 message to the
Support Service Provider (SSP)/Vendor, requesting them to find a solution.
However, the problems in these existing systems can be summarized as follow:
The size of the log in the CBE format is larger than that of the untranslated log.
(This drawback will reduce the system performance)
1
Call Home Format: This is the message transmission code between healing system and SSP/Vendor that
IBM&CISCO are undertaken for standardization. http://www.cisco.com/application/pdf/en/us/guest/
partners/partners/c644/ccmigration_09186a0080202dc7.pdf
374 J. Park et al.
• The disk, CPU and memory usage drastically increase in the process of
conversion, due to the complex calculations involved.
• The ASF has as many Adapters [9, 10] as there are components, and this may
cause a problem of insufficient resources, particularly in the case of handheld
devices used in ubiquitous environments.
• The ASF requires a high healing time, because immediate action time
corresponding to emergency situation is shortage.
• Furthermore, in the event that the component does not generate the log, it is
impossible for the system to heal itself.
3 Proposed System
The adaptation of proposed system is divided into three layers, such as Monitoring
Layer, Diagnosis & Decision Layer and Adaptation Layer. Fig. 1 shows an
architecture for self-healing, which is composed of the Monitoring Agent,
Diagnosis Agent, Decision Agent and Searching Agent. The proposed system
consists of 6 consecutive processes, viz. Monitoring, Filtering, Translation,
Analysis, Diagnosis and Decision.
Through resource status, log files and error events, if the Monitoring Agent detects
suspicious events of the components, it executes the CBE Parser in the Diagnosis &
Decision layer.
The first major attribute of a self-healing system is self-diagnosing [9]. The Diagnosis
Agent analyzes the CBE log, resource information (received from the Monitoring Agent)
and the dependency of the components, and then diagnoses the current problem (through
the Symptom DB). It provides technology to automatically diagnose problems from
observed symptoms. The results of the diagnosis can be used to trigger automated
reaction. Using the loging service existing in the operating system, as shown in Table1, it
classifies the Error Event, and sets up the priorities. The results of the diagnosis recognize
the situation level.
As shown in Fig 4, the Diagnosis Agent generates the Error Report and modifies the
CBE. The Error Report is an administrator document, and the CBE is a document for
the system. The following is the algorithm performed by the Diagnosis Agent Using
the first-order logic, we can recognize the situation level of system and represent the
policy for it. Fig. 5 illustrates context predicate and its example.
Through the information delivered by the Diagnosis Agent, The Decision Agent
determines the appropriate healing method with the help of the Policy DB. It also
receives feedback information from the administrator in order to apply the more
efficient healing method. The Information received from the Diagnosis Agent is used
to determine the healing method. The Decision Agent determines the solutions that
can be classified into root healing, temporary healing, first temporary healing and
second root healing. Temporary healing is a way of resolving a problem temporarily,
such as disconnecting a network connection, assigning temporary memory. The root
Self-management System Based on Self-healing Mechanism 377
The Table is the table to determine the optimal resolution method by analyzing
given attributes. Looking at the DECISION column, when placed under the current
diverse context, it helps to determine R (Root Solution), T (Temporary Solution), or
TR (first Temporary Solution, second Root Solution). The FEEDBACK Column is
showing feedbacks that were executed by the System Agent to heal the system. The
Decision Agent compares the fields with the information received by the System
Agent, these fields are CURRENT JOB, FUTURE JOB and AVAILABLE
MEMORY. If the value of the FEEDBACK Column is POSITIVE, the appropriate
method is determined.
It is used to search the vendor’s website for the knowledge required to solve the
problem. This Agent uses search engine (such as Google). It sends the resulting
search information to the administrator.
The Code Cache is used to provide healing code to solve the error of the component
arising in the emergency situations.
The Rule Model document identifies the different events that can be applied,
namely the “Emergency”, “Alert”, “Error” and “Warn” situation. Theses situations
have actions, linked to their respective situation, and then services of the proposed
system are reconfigured by
this strategy. We can
understand their behavior
from the above document.
If the agent classifies the
current situation as an
emergency situation, it acts
transformed code to heal
the component. The above
rule represented in XML is
transformed to:
the containers in JADE Agent Platform, and the ACL (Agent Communication
Language) is used to communicate among the agents. We performed the simulation
using six agents.
The Fig 9 shows extracted log data and resource monitoring for self-management
The proposed system was evaluated and compared qualitatively and quantitatively
in terms of the Log Monitoring Module, the Filtering & Translating Efficiency, and
the healing Time. (a) Log Monitoring Test. In the existing system, if the number of
components is , the system has to have processes to monitor the log. In the
proposed system, however, only one process is needed to monitor the log, as shown in
Fig. 10. In this figure, the proposed system demonstrates its ability to stay at a certain
level of memory usage, even when the number of components is increased.
380 J. Park et al.
(b) Filtering & Translation Efficiency Test. In the proposed system, the Component
Agent searches for a designated keyword (such as “not”, “reject”, “fail”, “error”, etc.)
in the log generated by the components. By using this approach, we were able to
increase the efficiency of the system, in terms of the size of the log and the number of
logs. We analyzed up to 500 logs, filtered out those logs not requiring any action to be
taken, and evaluated the number and size of the logs in the case of both the existing
and proposed systems. As a result of the filtering process, only about 20% of the logs
were required for the healing process, as shown in Fig. 10. Therefore, the proposed
system reduces the number and size of the logs, which require conversion to the CBE
format.
Fig. 10. Memory usage and comparison of size and number of logs
For each adaptation time, as shown in Fig.11, we verified that the proposed system’s
parsing time and healing time are fastest than the existing system’s, and rapidly
responded problems arising in the urgent situation. However, because the number of
monitoring factors is a little more, although the proposed system’s monitoring time
Self-management System Based on Self-healing Mechanism 381
was relatively a little load through providing the meaningful much more information
we verified that high quality of healing have been increased by monitoring
information. Although In the event that the error component does not generate log, we
couldn’t measure the healing time arising in the existing self-healing system because
the existing system was log-based healing system.
5 Conclusion
This paper has described self-management system for reliable system. The monitoring
layer consists of modules for monitoring the information such as log context,
resource, configuration parameters. Once monitoring module in the monitoring layer
detects an anomalous behavior and presumes that the behavior needs to be treated. In
the abnormal phase, modules in the diagnosis & decision layer are triggered. The
diagnosis & decision layer constitutes modules that filters, translates, analyzes,
diagnoses the problems, and decides its strategy. Finally, the adaptation layer
composes modules that execute the strategy selected in diagnosis & decision layer.
The advantages of this system are as follows. First, when prompt is required, the
system can make an immediate decision and respond right away. Second, the
Monitoring module monitors the generation of the log on the fly, thus improving the
memory usage. Third, before converting the log into the CBE (Common Base Event)
format, filtering is performed in order to minimize the memory and disk space used in
the conversion of the log. Fourth, it provides the fast healing time. Fifth, using the
Rule Model, the appropriate adaptation policy is selected. However, further decision
mechanism is likely to need to select the appropriate adaptation policy. Moreover this
approach may be extended for the intellignent sub-modules in the Diangosis &
Decision layer.
References
1. http://www.ibm.com/autonomic
2. IBM: Autonomic Computing: IBM's Perspective on the State of Information Technology,
http://www.ibm.com/industries/government/doc/content/resource/thought/278606109.html.
3. Jeffrey O. Kephart David M. Chess IBM Thomas J. Watson Research Center: The Vision
of Autonomic Computing, IEEE Computer Society, January (2003)
4. P. Oreizy et. al.: An Architecture-Based Approach to Self-Adaptive Software, IEEE
Intelligent Systems, Vol. 14, No. 3, May/June (1999) 54-62.
5. Garlan, D. and Schmerl, B.: Model-based Adaptation for Self-Healing Systems,
Proceedings of the First ACM SIGSOFT Workshop on Self-Healing Systems (WOSS),
South Carolina, November (2002) 27-32.
6. G. D. Abowd, R. Allen and D. Garlan.: Formalizing style to understand descriptions of
software architecture, ACM Transactions on Software Engineering and Methodology, Vol.
4, No. 4, October (1995) 319-364.
7. D. Batory and S. O'Malley: The Design and Implementation of Hierarchical Software
Systems with Reusable Components, ACM Transactions on Software Engineering and
Methodology, Vol. 1, No. 4, October (1992) 355-398.
382 J. Park et al.
Abstract. As the network technology has been developed, the Next Generation
Internet (NGI) such as Internet2, KOREN, KREONET2 and etc has been de-
ployed to support bandwidth of Giga bps. And, various applications such as the
video conference, the tele-surgery and etc that require high bandwidth has been
developed and operating on the NGI, especially KOREN and KREONET2 in
Korea. When the applications are operating and happen to face some problems,
various tools for traffic measurement are usually used to analyze them, and pro-
vide traffic information in each link on total flow path or other metrics such as
end-to-end delay, jitter and etc in most cases. However, although with these
tools most of network problems can be properly analyzed, the problem in the
user's view point can not be resolved so that which part of networks and/or
which ones among user's systems and nodes on the flow path cause the prob-
lems for the user's flow is not discriminated Therefore, this paper designs a end-
to-end performance monitoring system for the user's flow that the user can ac-
cess performance data for user's flow on the flow route, and describes the ex-
perience about deployment on KOREN.
1 Introduction
As the network technology has been developed, the Next Generation Research Net-
works have been deployed to support bandwidth of Giga bps. As the international re-
search network such as GLORIAD[3] as well as the domestic research network such
as Internet2 in U.S.A.[1] and KOREN[2] in Korea appear, there have been various
collaborations for research and performance through the network. Especially, many
activities such as the video conference, the tele-surgery and etc through KOREN and
KREONET2 have been internationally performed by using applications that transport
the high quality multimedia data of the high-definition (HD) level in real time.
When experiencing these activities to use the applications, various problems are of-
ten faced, but usually the end users should realize that they have no special means
without discussing the phenomena with the end users on the opposite site. For exam-
ples, if you hear noise or watch the distorted video stream from the network, you can
*
This work has been supported by NCA, Korea.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 383 – 392, 2006.
© Springer-Verlag Berlin Heidelberg 2006
384 W.-C. Song and D.-J. Choi
not discover whether it comes from the network problems or from system itself on ei-
ther site. Therefore, if information about users’ flow can be provided to the users
themselves as a way to solve these problems in network events, we think that they can
as least understand whether the source of the faced phenomena is network problems
or not.
As many open source tools have been already developed to investigate the network
situations, we can easily get very various kinds of information from networks, but
most of them do not consider the user’s individual flow information but just show
current traffic amount and available bandwidth on the path or some link. Although
much bandwidth is available on the flow path, we have experienced in many cases
that users can not use fully the bandwidth if the end systems are not tuned properly.
So, enough bandwidth means neither the user’s flow can use it nor the user currently
uses it. And, the network operator can collect useful information for flows from every
router, but as it is provided specially depending upon the user’s request, users can not
easily access user’s individual flow information.
Therefore, this paper designs and deploys the end-to-end performance monitoring
system on KOREN, so that users can easily get information about their flow’s infor-
mation in the user’s viewpoints and analyze their communication status by them-
selves. This paper consists of four sections: after introduction, we explain the system
architecture in section 2, and describe the deployment on KOREN in section 3. Fi-
nally, section 4 concludes and describes the further research.
This architecture has been already taken by NeTraMet [6], and IPFIX [7] also takes
the similar approach. With this basic architecture, we intend to design the network-
wide system that each meter on routers collects the flow information from its router
and transfer it to the centralized reader so that the reader can analyze it and provide
useful flow information to the users. Therefore, we use the meter as the original con-
cept and extend the reader to the centralized one that decides/coordinates to get which
flow information from which router on the path. The manager must be designed, but
as it is related to the router’s security issue and it practically belongs to network op-
erator’s authority, design of manager is not considered in this paper.
2.1 Meter
Meter is a component that measures flow traffic in a node, and is RTFM itself. Actu-
ally, as this component should be operating on a router, it is necessary to be depend-
ent on the router machine. So, it may be reasonable to select a tool that many routers
can have on themselves. As Netflow [4] is well-known as one of the best tools for
measuring flow information, and can be deployed on the router of many vendors, we
select it as the meter in the design.
2.2 Reader
As we state earlier, in this paper we extend the reader concept of RTFM to the cen-
tralized network-wide Reader. It should have role to identify the requested flow’s dy-
namic path, request the flow information to the corresponding routers (meters), and
provide the client the received flow information from routers. For the purpose, the
reader is designed as shown in the figure 2 that consists of the Request Handler, the
Manager, the Coordinator, the Netflow data DB, and the Dynamic Flow Path Detec-
tion modules. And, the clients in the figure 2 are entities that are now generating the
traffic in the network and want to get their own flow information from Reader.
386 W.-C. Song and D.-J. Choi
information directly to the meter in the KOREN routers, this investigation extracts the
flow information hop-by-hop from the DB. As an example, figure 4 describes that
when user systems connected respectively to Seoul node and Gwangju node have
video conference by using DVTS[9], the Netflow data collected from Seoul node
have the IP address of Daejeon node as the Next hop router IP address value. So,
the total flow path can be hop-by-hop searched as Seoul->Daejeon->Gwangju.
2.3 Client
The Client’s functions are very simple. The user who wants to request the flow infor-
mation fills out about the source IP/port, destination IP/port and update duration
items. Then, the Client sends the above items with current time to the Reader and gets
the flow information from the Reader repeatedly in the update duration. And, the
server application port is well known in most application, but as the client applica-
tion’s port is the ephemeral port number, if a port number is not filled in the request,
the Client investigates the port number by using the given IP and port numbers from
client system’s OS. Figure 5 shows the Client GUI that replies the flow information in
text.
We have developed the client as the client/server model, not the web-based
model. It is because we consider development of the visualization part in the fu-
ture. If the system is developed in the web-based model, the server (the Reader)
may have the computing burden for visualizing the data. So, we have thought that
it would be better the Client receives the flow information as the numerical data
and computes the visualization. The computing burden for visualization shall be-
long to the client machine.
Experiences in End-to-End Performance Monitoring on KOREN 389
3 Deployment on KOREN
While developing the system, we have faced some limitations in the Netflow although
the Netflow has many advantages over the other utilities. First, the Netflow performs
1/10 sampling to generate the flow information in KOREN. Although we have under-
stood that sampling is necessary in order to maintain proper performance of the router
in WAN environment, sampling may cause the uncertainty of the flow information to
increase. In applications such as DVTS to require high bandwidth, the flow informa-
tion from Netflow looks still valid, but applications with low data rate can not be de-
tected in the flow information in some cases. Therefore, the system designed in the
paper can be used only for flows requiring high bandwidth such as HD-quality multi-
media stream.
Second, Netflow can generate its data only in one minute interval at most. It means
the minimum interval to generate data is one minutes and it can not timely generate
data. It causes synchronization problems among Netflow data from several KOREN
routers. If someone want to get his/her flow information at some time, it might be
failed because routers can not generate the information timely. Therefore, the flow
data provided to users should be at most one-minute earlier data. However, we think
this system is still valid because in many cases it may take one or more minutes for
users to start monitoring the flow after recognizing the problems, and in some cases
users may want to understand only the trends about the traffic.
Third, too many Insert DB commands are required to collect the Netflow data into
the DB. At one time, the number of packets received from KOREN was 84 during one
minute. The length of each Netflow packet is 1,464 bytes, and the storing unit for a
flow in the Netflow data is 48 bytes. Therefore, collecting the Netflow data into DB
requires 156,990 Insert commands as follows:
Experiences in End-to-End Performance Monitoring on KOREN 391
4 Conclusions
The end-to-end flow monitoring system designed in this paper provides users their
own individual flow information in the user’s viewpoints. We use the RTFM as the
basic architecture and extend it to the network wide system. Our system mainly redes-
igns the Reader part to investigate the flow path and get the flow information from the
corresponding routers on the path. We think this system could be very useful to users
who want to use applications requiring high bandwidth on networks. If they face
some problems, we think they can easily investigate the cause, analyze it, and request
help to the NOC with details about the problems. We have adopted the Netflow as the
meter in the system. Due to some limitations of the Netflow we have faced some
problems in deployment, but they have been successfully resolved. In the future, we
will correlate the flow information supported by this system with traffic information
measured by SNMP, and design the visualization effectively to provide very useful
information to users.
References
1. http://www.internet2.edu/
2. http://www.koren21.net/
3. http://www.gloriad.org/
4. Cisco System, "NetFlow Services and Applications," White Papers, http://www.cisco.
com/warp/public/cc/pd/iosw/ioft/neflct/tech/napps_wp.htm
5. Brownlee, N., Mills, C. and G. Ruth, "Traffic Flow Measurement: Architecture", RFC
2722, October 1999.
392 W.-C. Song and D.-J. Choi
1 Introduction
Recently as new network technologies continue to make appearance in the market,
there is a fast growth in converged products as they combine with the existing
products. WiBro service will be launched as a Fixed Mobile Convergence product,
while Home Networking service where various home information devices converge
into a single network and IP TV that allows TV viewing through broadband Internet
connection are being introduced.
In line with the launching of such convergence service products, Operation Support
System (OSS) must also take on architecture where appropriate management
functions can be accommodated at the right time. Still, as the OSS becomes
complicated due to the addition of numerous functionalities, it often becomes more
difficult to accommodate the management function. In this case, modification of even
a single function in the system may cause a reverberating effect on other components,
thus, the whole of chain-reaction must be taken into account upon making a
modification. This process of adding or modifying a function in the OSS costs a lot of
money, not to mention the difficulty in identifying the root cause of a problem when
fault occurs after the launch. If legacy components have a high degree of inter-
dependency with other components, the reusability of those existing components
declines. In case of a convergence service product, due to the fact that it is a
combination of several disparate services, it may be much more efficient, in terms of
the saving of cost and time, to develop it by utilizing functionalities for the existing
services. But if re-utilizing existing components is not easy, provisioning of service at
the right time can become a tall task.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 393 – 402, 2006.
© Springer-Verlag Berlin Heidelberg 2006
394 Y.-W. Woo et al.
The most effective solution for this problem may be the Service Oriented
Architecture (SOA)[1]. The basic component of SOA, service, can be loosely coupled
with each other, thus weakening the relationship between components and
implementing OSS by combining modularized services are possible. This way, the
system can quickly respond to the addition of management functionalities for the
convergence service or new functionalities for existing services. Also, as services
become modularized, maintenance will become easy, and with the standardization of
service interface, they will have interoperability with other modules that were
developed into different platforms[2]. And, the concept of SOA can be a valuable tool
for Implementing NGOSS(Next Generation Operation System and Software) program
which is suggested by TM Forum.
This paper proposes next generation SOA-based OSS architecture to implement
telecom services. After closely examining the architectural aspect of the service unit
of the OSS and searching for the most appropriate architecture for processing them,
the design methodology and OSS architecture for services, which are the basic unit in
SOA configuration is proposed In Section 2, we describe the requirements about
NGOSS. And, we introduce the SOA as the OSS architecture which satisfy parts of
the requirements for NGOSS. In Section 4, we suggest the method to deduce services
from operations from OSS, and suggest next generation OSS architecture by
describing constituent part of it. In Section 5, we will represent next generation OSS
architecture is applicable to convergence service by giving an example of the WiBro
service.
Web service[6] is the leading technology that allows such SOA service
implementation. Web service is comprised of standardized technologies independent
of platform, such as Universal Description, Discovery and Integration (UDDI)[7],
Web Services Description Language (WSDL)[8], and Simple Object Access Protocol
(SOAP)[9]. Each of these has the following roles in service implementation: UDDI
registers web service and provides mechanism for finding/binding which corresponds
to the Service Registry in figure 1; WSDL is used to define the format of the interface
396 Y.-W. Woo et al.
and input/output message for using the web service whose role is to contain the
details of services in the Service Registry of figure 1; and SOAP is the SML protocol
standard that defines the message format used in web service request and response.
For the implementation of a process that is a coupling of atomic services or
composite services, an integrating layer is required. For this, Business Process
Management (BPM)[10], Enterprise Service Bus (ESB) and Business Activity
Monitoring(BAM) are used. Figure 2 shows a system architecture that includes all of
these. BPM is run as a management system and tool for process automation and
management as it takes charge of the system implementation with service as the basic
unit for business logic implementation. For efficient integration of distributed service
components, ESB acts as a middleware that provides process-based integrated support
and support for web service-based integrated environment. Aforementioned services
are implemented based on consolidated database, and are coupled and orchestrated.
Composite Service, which is a collection of orchestrated fundamental services, can
also be used as a fundamental service for another round of orchestration. BAM plays
a role of monitoring business processes which is developed by BPM and visualized
the present condition of business activities of enterprise. Here, BPM and ESB can be
considered as means that actually implement the concept of ‘Externalized Process
Control ’ and ‘Common Communication Vehicle’ of NGOSS.
To incorporate business logic in a system, the key issue is how to design service, the
fundamental unit. Through the analysis of the functional and architectural aspects of
services and the review of the performance aspect, the following criteria for
decomposing services can be derived.
SOA-Based Next Generation OSS Architecture 397
If service binding time ‘b’, represents the total time it takes when a service calls to
locate another service, binds it and parses the SOAP header, and if Level n service is
designed to call Level n+1 services m times, the total time it takes can be calculated
as follows:
SOA-Based Next Generation OSS Architecture 399
m
Tlevel (n ) = b × ¦ (Tlevel ( n+1) ) (1)
x =1
If Level n service is composed of a number of Level n+1 services, the time it takes
to execute Level n service will increase as the number of service decomposition levels
increases. Figure 4 illustrates this in a line chart. If the number of service
decomposition level is increased, the frequency of recurrent calling of sub-level
services will increase, dropping the performance. Therefore, in the implementation
step, it has to be decided that up to which level of decomposition shall be
implemented as a service unit. The decision can be made based on the number of
services, time required to call a service, and processing time of each service.
After every operation and service of the system are defined, the data model and
repository is needed to contain the communicated data among them. The data can be
modeled by analyzing operations from the service map and deriving entities and
attributes from them. NGOSS suggests SID(shared Information and Data model) as
data model for the common repository. It represents that whole services in the system
should have shared information and data model and should not have the different data
repository to deal with the same information. It can be obtained by building
consolidated database, and all the services in the system should communicate each
other by using the data from it.
In order to manage the services derived from the service map efficiently, Service
Registry Management functionalities are required. It provides the functions such as
400 Y.-W. Woo et al.
registering, updating and deleting the services to the service registry. It also provide
authority management mechanism that keeps unauthorized user or system from
finding and executing the important services. And, it need to keep an eye on whether
the services is executed properly or not and log the result when the error occur. It also
need to track and testing the execution time of each services and log the execution
time of them. These logged results help system or operator making plans to use the
services. And It also need to log user list and history for each services in order to
prevent the services from being executed by too many systems and users at once.
If the functionalities to manage services are prepared, the business process can be
developed by orchestration of services securely and efficiently. Figure 5 shows the
SOA-based Next Generation OSS architecture which contains all the described
features previously.
WiBro requires the management of both fixed and wireless sections as shown in
the network topology. When a resource trouble occurs, operators shall be able to
locate in the assurance management view in which section the trouble occurred. Table
2 shows a sample of decomposed services of WiBro derived by using the
aforementioned service map. Here, the service of ‘1.1.3.PD.DI Testing Resource
Trouble’ may be decomposed into two lower level services: device test functions for
the fixed section and the wireless section, and then further be decomposed into test
functions by equipment type in each section and services can be orchestrated and
developed with the architecture which is shown in figure.5.
During this process, if a testing function for the fixed section is already registered
as a web service at the Service Registry, the test function module can be used during
assurance management of WiBro. That is, a simple addition of a device testing
function for the wireless section will be sufficient, so cost as well as development
time will be reduced.
Also, by publishing the testing function as a web service, this function can be shared
by other systems that also need such function. For example, as demonstrated in figure 7,
it is assumed that an existing network management system in charge of Metro-Ethernet
makes a device testing function for L2 switch and publishes it to the Service Registry. If
WiBro service is implemented afterwards, in developing WiBro NMS, the desired Metro
Switch testing function can be searched in the Service Registry for use. Other systems
can always make use of this service from the Service Registry.
6 Conclusion
This paper proposed SOA-based Next Generation OSS architecture which satisfies
NGOSS key principles from TM Forum by applying the basic concepts of SOA. We
suggested the service map which is the tool to derive services from operations and we
described BPM which orchestrate services and develop business process, consolidated
database which provides common communication language between each services,
service registry manager which enables services can be used properly. And, this paper
also presents adopting new convergence service will be easier than legacy system by
reusing existing service with service repository with example of WiBro.
References
1. Roy W. Schulte Yefim V. Natis, “Service Oriented Architecture” Gartner Group, SSA
Research Note SPA-401-068,1996
2. Randy Heffner, “The Big Strategic Impact Of Organic Business And Service-Oriented
Architecture, Jun 2004
3. TMF, “Enhanced Telecom Operation Map (eTOM) – The Business Process Framework”,
GB921, V6.1 Nov 2005
4. TMF, “Shared Information and Data Model (SID), GB922, V6.1 Nov 2005
5. TMF, “The NGOSS Technology Neutral Architecture”, TMF053 (and Addenda), Aug
2004
6. W3C Web Services WG, “Web Services Architecture”, http://www.w3.org/TR/2004/
NOTE-ws-arch-20040211/, Feb 2004
7. www.uddi.org, “UDDI Technical White Paper”, Oct 2004
8. David Booth,Canyang Kevin Liu, “Web Services Description Language(WSDL) Version
2.0 Part 0: Primer”, http://www.w3.org/TR/2006/CR-wsdl20-primer-20060327, Mar 2006
9. Nilo Mitra, “SOAP Version 1.2 Part 0:Primer”, http://www.w3.org/TR/2003/REC-soap12-
part0-20030624/, Jun 2003
10. Howard Smith, Peter Finger, “Business Process Management the third wave”, Meghan-
Kiffer press, 2003
Performance Analysis of a Centralized Resource
Allocation Mechanism for Time-Slotted OBS Networks
Tai-Won Um1, Jun Kyun Choi2, Seong Gon Choi3, and Won Ryu1
1
Electronics and Telecommunications Research Institute,
161, Gajeong-dong, Yuseong-gu, Deajeon, 305-350, Korea
{twum, wlyu}@etri.re.kr
2
Information and Communications University,
P.O. Box 77, Yusong, Daejon, 305-348, Korea
[email protected]
3
Chungbuk National University
12 Gaeshin-dong, Heungduk-gu, Chungbuk, Korea.
[email protected]
1 Introduction
Optical burst switching (OBS) [1], [2] is a novel optical switching paradigm, which is
capable of enhancing optical channel utilization by multiplexing collected packets,
referred to as an optical burst, onto a wavelength. There have been a number of OBS
reservation protocols, such as just-enough-time (JET) [1], just-in-time (JIT) [3],
Horizon, and wavelength-routed OBS (WR-OBS) [4], [5].
In JET, an ingress node transmits a control packet through a dedicated control
channel in order to configure each OBS switch for the burst duration along the paths
before sending out its corresponding optical burst. The control packet and optical
burst are separated by a certain delta time called offset, which is required because the
control packet is electronically processed at intermediate switches while the optical
burst is not. The OBS inevitably suffers from an optical burst loss problem due to a
contention for obtaining optical resources between control packets within an
intermediate node if the number of burst simultaneously arrived at an output port of
an OBS switch is more than the number of available channel.
In WR-OBS, while an ingress node aggregates packets into a burst, it sends a
wavelength request to a control node. When an acknowledgement is received, the
burst is assigned to an available wavelength-routed path. Compared to conventional
JET, this scheme is able to support explicit QoS provisioning and deterministic delay
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 403 – 411, 2006.
© Springer-Verlag Berlin Heidelberg 2006
404 T.-W. Um et al.
for the optical burst, but it wastes wavelength resource during signaling because
intermediate nodes begin reserving the whole wavelength for that request when the
acknowledgement arrives there. Moreover, as the burst size decreases and signaling
delay for establishing a wavelength-routed path increases, channel utilization
decreases.
On the other hand, Time Slotted Optical Burst Switching (TS-OBS) [6],[7] is a
form of OBS where time is essentially quantized into discrete units, referred to as
slots. Obviously, in TS-OBS based on synchronous transmission the chance of
contention is smaller than OBS based on asynchronous transmission because the
behavior of the bursts is more predictable and regulated [8].
This paper describes a slot assignment procedure using centralized control for TS-
OBS networks. In this scheme the controller is responsible for finding a path,
configuring time-slotted optical cross connects (ts-OXCs) [9],[10] and assigning a
time-slot to transmit an optical burst. Our scheme intends to improve burst contention
resolution and optical channel utilization. The rest of the paper is organized as
follows. In Section 2, we describe the TS-OBS network architecture with centralized
control, and Section 3 focuses on the centralized resource allocation mechanism. The
results obtained from analysis and simulations will be discussed in Section 4. Finally,
we draw our conclusions in Section 5.
2 Network Architecture
Network control methods can be classified into three categories: centralized,
distributed and hybrid control. Each control method have pros and cons, but typical
telecommunication networks and automatic switched optical networks (ASON) [11]
in ITU-T follow the centralized control architecture, in that the control plane is
separated from the data plane.
In this paper, we consider the centralized control architecture for TS-OBS
networks in order to execute the path calculation and slot assignment. As shown in
Fig 1, the control plane consisted of controllers is separated from the data plane
consisted of ts-OXC, and there is an interface channel between a controller and a ts-
OXC, which can communicate each other by defining signaling messages.
In this TS-OBS network, ts-OXC is made up of all-optical switching components,
operates on a synchronized timing basis without O/E/O conversion. The wavelengths
on the optical links are split into time-slotted wavelengths of a fixed-size divided by a
time interval. One or more time-slot(s) can be assigned for the request from the
ingress ts-OXC to transmit an optical burst. All the link propagation delays are an
integer number of time slots. On the path from the ingress ts-OXC to the egress ts-
OXC, each ts-OXC has a role in switching incoming time-slots to outgoing time-slots
in order to transport the optical burst. In order to do this, every ts-OXC needs to
maintain switching tables containing information entries that are received from the
controller, to forward slotted wavelengths.
The network topology information will be integrated to the controller by routing
protocols. By applying the time-slot assignment mechanism into the routing table, the
controller may find a number of paths. One path could be selected according to
constraints given for the request.
Performance Analysis of a Centralized Resource Allocation Mechanism 405
4 Performance Analysis
To determine the channel utilization for our proposed scheme, we extend the
analytical model introduced in [5] for WR-OBS networks. For clarity and simplicity,
the analysis in this section is based only on mean values for all parameters.
Performance Analysis of a Centralized Resource Allocation Mechanism 407
The burst aggregating cycle can be explained as follows. Let the ingress delay
tingress be the burst aggregation time. We define tprop,control to be the propagation delay
for the control packet. Processing the time-slot request requires time tproc, followed by
an acknowledgment packet to be returned to the requesting ingress node, with an
additional delay tprop,ack. After that, a wavelength channel is allocated during slot time
tslot. Actually, there is a guard-time (tguard) between the time-slotted wavelengths
because each intermediate optical switch requires switching time and a slot can be
interfered with by an adjacent slot by dispersion and non-linear effects on the physical
layer. So, tWHT consists of tslot and tguard. In slot time tslot , a burst in the buffer of an
ingress ts-OXC is sent. Let the bit rate bin be the aggregated bit rate for the traffic
from all n sources directed to a particular destination and requiring the same QoS.
Bursts are transmitted from the queue at core bit rate bcore, where bcore > bin. The
number of wavelengths is A = bcore / bin. This time is related with ttrans = Lburst/bcore,
which is the time to complete the burst transmission. We assume that tslot and ttrans are
the same to simplify this analysis. The propagation delay, tprop, is a delay to propagate
bursts for receiving an egress ts-OXC.
The maximum deterministic latency or upper bound on the maximum transmission
time that packets experience between entering the core network at the ingress ts-OXC
and egress ts-OXC is
Latencymax = tingress + tWHT + t prop
.
The arrival of the acknowledgment packet from the controller sets the start of the
subsequent burst aggregation and the cycle repeats. From the analysis of the timing
involved in burst aggregation and transmission, it is clear that network efficiency
depends on the processing speed of the network controller.
It is assumed that burst sizes increase linearly, equivalent to the case of CBR traffic
arriving to the buffer. Thus, for a constant load and CBR traffic, the burst size Lburst is
proportional to the ingress delay and the input bit-rate bin, so that Lburst = bin · tingress.
A parameter following from [5] is the bandwidth per wavelength, which indicates
the effective bandwidth of a lightpath used for transmission of data between edge ts-
OXCs.
Lburst bcore ∗ tslot
B per slot = =
tWHT t guard + tslot
The influence of the guard time is shown in Fig. 3(a) for bcore = 3 Gbps, and tguard =
0.15 ms, 1.5 ms, 15 ms. The increase in bandwidth for the identical values tslot is
reduced for higher tguard; for tguard = 15 ms, values remain below 2.6 Gbps for a 3 Gbps
physical bit rate. Fig. 3(b) shows the effect of bandwidth saturation for tguard = 1.5 ms
and core bit rates vary from 1 to 5 Gbps. The significance of the results is that Bper
slot remains significantly smaller than the optical line rate for tslot 40/A ms, because
the packet loss probability is too high to send the burst when tingress is bigger than
40ms.
408 T.-W. Um et al.
Fig. 3. Bandwidth per slot Bper slot for (a) bcore = 3 Gbps, and tguard = 0.15ms, 1.5ms, 15ms and
for (b) tguard = 1.5ms, bcore = 1 Gbps, 3 Gbps, 5 Gbps
The network topology and parameters for the simulation are given in Fig. 4. We
consider 5 ingress ts-OXCs, 1 core ts-OXC and 1 egress ts-OXC, which support full
wavelength conversion. We assume that the average arrival rate is the same for all
Performance Analysis of a Centralized Resource Allocation Mechanism 409
ingress nodes. Packets arrive at each ingress node according to a Poisson process with
a 2Gbps input bit rate. Packets are aggregated into a burst of 12.5kbyte in size at the
ingress node, and sent to the egress ts-OXC. The performance metrics are ingress
queueing delay, buffer size and link utilization as a function of offered input traffic
load.
Fig. 5 shows the link utilization of our TS-OBS scheme using a centralized slot
assignment as a function of the offered traffic load per ingress node. Comparing the
link utilization of conventional OBS with the proposed scheme shows that our scheme
improves the utilization markedly. In conventional OBS, if optical bursts collide with
one another, they will be dropped. However, in our scheme, by sending the control
packet to the centralized controller, the ingress ts-OXC can prove the available time-
slots, and if there are no available time-slots, the ingress node will try to reserve the
next slot again until it succeeds, instead of dropping the optical burst.
Fig. 6 shows the ingress buffering delay versus the offered load when the
propagation delay from the ingress node to the controller is 2ms, 4ms or 6ms. In the
proposed scheme, if there is not an available time-slot, the buffered data will wait
until it reserves a time-slot. The signaling time to reserve a time-slot from the ingress
node to the controller depends on the distance between them. Therefore if the
propagation delay to the controller increases, the assembled packets experience more
ingress delay, and the ingress node needs more buffers (as shown in Fig. 7).
In conventional OBS, when a burst is blocked, the only way to recover the burst is
through TCP retransmission. In order to achieve a connection oriented session, TCP
makes use of retransmission on timeouts and positive acknowledgments upon receipt
of information. However, TCP does not provide fast recovery due to its host-to-host
behavior and time-out mechanism. On the other hand, UDP is connectionless, which
means that it can not provide error control and flow control. Therefore, if we consider
the TCP/UDP layer’s retransmission of lost packets in conventional OBS, buffering at
the OBS layer in our scheme may support better performance for upper layers. In this
paper, we do not analyze the performance of the TCP layer; that remains for future
studies.
1.0 0.020
0.9 0.018
AVERAGE INGRESS DELAY (s)
0.8 0.016
AVERAGE UTILIZATION
0.7 0.014
0.6 0.012
0.5 0.010
0.4 0.008
0.3 0.006
0.2 0.004
0.0 0.2 0.4 0.6 0.8 0.0 0.2 0.4 0.6 0.8 1.0
Fig. 5. Offered load vs. Link utilization Fig. 6. Offered load vs. Ingress buffering delay
410 T.-W. Um et al.
4.0M
3.5M
2.5M
2.0M
1.5M
1.0M
500.0k
Ingress-to-controller: 2ms
Ingress-to-controller: 4ms
Ingress-to-controller: 6ms
0.0
0.0 0.2 0.4 0.6 0.8 1.0
Offered Load
5 Conclusion
In this paper, we proposed a centralized control architecture and time-slot assignment
procedure for TS-OBS networks. In this architecture, a request of path calculation and
slot assignment from an ingress ts-OXC is delivered to a controller. Upon receiving
the request, the controller takes the responsibility of finding an optimal path,
delivering the configuration information to ts-OXCs and informing the assigned time-
slot to the ingress ts-OXC. Analysis and simulation results show that the channel
utilization of the TS-OBS network is improved markedly at the expense of ingress
buffering delay.
Acknowledgments. This work was supported in part by the Korea Science and
Engineering Foundation (KOSEF) through the Ministry of Science and Technology
(MOST); and the Institute of Information Technology Assessment (IITA) through the
Ministry of Information and Communication (MIC), Korea.
References
1. M. Yoo, C. Qiao and S. Dixit, “QoS Performance of Optical Burst Switching in IP-Over-
WDM Networks,” IEEE JSAC, vol. 18, no. 10, pp. 2062-2071, Oct. 2000.
2. C. Qiao, “Labeled Optical Burst Switching for IP-over-WDM Integration,'' IEEE
Communication magazine, vol. 38, no. 9, pp. 104-114, Sep. 2002.
3. J. Y. Wei and R. I. McFarland, “Just-In-Time Signaling for WDM Optical Burst Switching
Networks,” IEEE JLT, vol. 18, no. 12, pp. 2109–2031, Dec. 2000.
4. Michael Duser, Polina Bayvel, “Performance of a Dynamically Wavelength-Routed
Optical Burst Switched Network”, IEEE Photonics technologies letter, vol. 14, no. 2, pp.
239-241. Feb. 2002.
5. Michael Duser, Polina Bayvel, “Analysis of a Dynamically Wavelength-Routed Optical
Burst Switched Network Architecture”, Journal of lightwave technologies, vol. 20, no 4,
pp. 573-585. Apr. 2002.
6. Jeyashankher Ramamirtham and Jonathan Turner, “Time Sliced Optical Burst Switching,”
IEEE INFOCOM’03, pp. 2030-2038, Apr. 2003.
Performance Analysis of a Centralized Resource Allocation Mechanism 411
7. Geoffrey M. Garner, “Advanced Plesiochronous TSOBS,” prepared for SAIT, Dec. 2004.
8. S. Yao, S. J. B. Yoo and B. Mukherjee, “A comparison study between slotted and
unslotted all optical packet-switched network with priority-based routing,” OFC’01, Mar.
2001.
9. I.P. Kaminow, et al., “A Wideband All-Optical WDM Network”, IEEE Journal on
Selected Areas in Communications. Jun. 1996.
10. Nen-Fu Huang, et al., “A Novel All-Optical Transport Network with Time-Shared
Wavelength Channels”, IEEE Journal on Selected Areas in Communications, Oct. 2000.
11. “Architecture for the automatically switched optical network (ASON)”, ITU-T
Recommendation, G.8080/Y.1304, Nov. 2001.
12. Bo Wen, Krishna M. Sivalingam, “Routing, Wavelength and Time-Slot Assignment in
Time Division Multiplexed Wavelength-Routed Optical WDM Networks,” IEEE
INFORCOM 2002.
13. Panos Trimintzios, et al., “A Management and Control Architecture for Providing IP
Differentiated Services in MPLS-Based Networks,” IEEE Comm. Mag., May 2001.
14. Jun Kyun Choi, et al., “Fast Protection Mechanism Using Centralized Control with
SRLG”, IETF draft, draft-choi-centralized-protection-00.txt, Jul. 2004.
Efficient Performance Management of
Subcarrier-Allocation Systems in Orthogonal
Frequency-Division Multiple Access Networks
1 Introduction
Next-generation wireless networks even provide high capacity and high data-rate
services, enabling high quality images and video to be transmitted and received.
Orthogonal frequency-division multiplexing (OFDM) is a potential key technology
for the next-generation broadband wireless networks [1,2]. It has several advantages,
such as flexibility of allocating subcarriers to users, high spectral efficiency, low
receiver complexity and simple implementation by inverse fast Fourier transform
(IFFT) and FFT [3,4].
Subcarriers in the orthogonal frequency-division multiple access (OFDMA)
networks are scarce and valuable, so operators should employ them efficiently by the
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 412 – 421, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Efficient Performance Management of Subcarrier-Allocation Systems 413
user-request data rate. The available bandwidth in a base station is limited and shared
if the values of AMC parameters are decided and used. Every user shares the FFT
uplink space, and the base station assigns subcarriers to users [3]. Without loss of
generality, assume that a cell (a base station may contain one or a few cells) owns a
subcarrier-allocation system.
λn
λn
λ3
…
λ2
λ1 λ2
λ1
… m m+1 m+2 m+3 … m+n …
mμ (m+1)μ
Fig. 2. Demonstration of the flow-in flow-out diagram for state m on the OFDMA subcarrier-
allocation system
λn
λn λmin( m ,n ) λmin(c−m ,n )
…
…
λ3
λ3
…
…
λ2 λ2
λ2 λ2
λ1 λ λ1 λ1
0 … …1 m … … c
μ mμ (m+1)μ cμ
Moreover, the system has no buffer (the queue length equals zero), making it a loss
system. Thus, the subcarrier-allocation system can be modeled on the batch-arrival
model MX/M/c/c, whose state-transition-rate diagram is shown in Fig. 3, where 1 ≤ m
n
≤ c − 1. If c ≥ n, then state 0 has the flow-out rate λ = ¦ λk . When m < n, the
k =1
maximum batch-arrival size from some previous state is given by m. Thus, the
maximum rate of incoming arrivals into state m should be λmin(m,n). If m ≤ c − n, then
all outgoing arrivals from state m have the same case, as illustrated in Fig. 2.
However, the arrivals only have partial success in being served with smaller size than
c − m + 1 when c − n < m ≤ c – 1. Hence, the outgoing arrival rates from state m are
given by λmin(c−m,n).
By queueing theory, the equilibrium (steady-state) equations written below are run
to obtain the steady-state probabilities pm of the model.
n
λp0 = μp1 , where 1 ≤ n ≤ c and λ = ¦ λk . (1)
k =1
min(c − m, n ) min( m, n )
§ ·
¨¨ mμ + ¦ λk ¸¸ p m = ¦ λk p m−k + (m + 1)μp m +1 , where 1 ≤ m ≤ c − 1 . (2)
© k =1 ¹ k =1
416 J.-C. Chen and W.-S.E. Chen
n
cμp c = ¦ λk pc −k , (3)
k =1
which can be used for verification. Reforming (1) and (2) yields
p1 = p 0 λ μ , and (4)
min(c − m, n)
ª§ · min(m, n )
º
pm+1 = «¨¨ mμ + ¦ λk ¸¸ p m − ¦ λk p m−k » (m + 1)μ , where 1 ≤ m ≤ c − 1 . (5)
¬© k =1 ¹ k =1 ¼
Recursive programs cannot always solve the equations, owing to overabundant
recursive levels for large c. Therefore, an iterative procedure is adopted to solve the
equilibrium equations as follows.
Let p 0 = 1; then p1* = p 0* (λ μ ) = λ μ .
*
(6)
min( c − m , n )
ª§ · min( m , n )
º
p m* +1 = «¨¨ mμ + ¦ λk ¸¸ p m* − ¦ λk p m* −k » (m + 1)μ , where 1 ≤ m ≤ c − 1 . (7)
¬© k =1 ¹ k =1 ¼
c
According to the normalizing condition ¦p i = 1 , the equilibrium probabilities of
i =0
Once the equilibrium state probabilities are known, the CBP and BU can be
derived forwardly. If a new call finds that the available capacity in the corresponding
cell cannot meet its rate requirement, then it is fully blocked. CBP is expressed as the
number of blocked calls divided by the number of total calls during a period of time.
Thus the CBP of the batch-arrival model can be written as
n −1 n n
§ ·
Ω = ¦ ¨ p c −i ¦ λ k λ ¸, where 1 ≤ n ≤ c, and λ = ¦ λk . (9)
i =0 © k = i +1 ¹ k =1
Ls equals the mean number of busy servers in the system, because the queue size is
zero. Observing the system for a long period of time T, we have the average BU
shown below.
Efficient Performance Management of Subcarrier-Allocation Systems 417
n c n c
ª 1º ª º
¦ «λ T (1 − ¦ p ) ⋅ k ⋅ μ »
k i L'Hospital ¦ «¬kλ (1 − ¦ p )»¼
k i
Ψ1 = lim
k =1 ¬ i =c −k +1 ¼ = k =1 i =c −k +1 , (11)
T →∞ T ⋅c cμ
n c
ª º
where ¦ «¬kλ
k =1
k (1 − ¦ p )»¼ = gˆλ
i =c − k +1
i eff ; 1 ≤ n ≤ c, and 1/μ is the mean call holding time.
Period T has λk T calls with kRb incoming, where n ≥ k ≥ 1, and thus a total of
c
λ k T (1 − ¦p ) i rate-k calls are served. The rate-k bandwidth usage is given by
i =c − k +1
c n c
ª 1º
λ k T (1 − ¦ p ) ⋅ k . Therefore, ¦ «λk T (1 − ¦ pi ) ⋅ k ⋅ μ »
i represents the total usage
i = c − k +1 ¬ k =1 ¼
i = c − k +1
time of all possible successful users in the period T, where 1/μ is the mean call
holding time. The numerical analysis demonstrates that
Ψ2 = L s / c = Ψ1 . (12)
0.9
0.8
0.7
0.6
CBP/BU
0.5
0.4
0.3 CBP-Theoretical
CBP-Simulation
0.2 BU-Theoretical
0.1 BU-Simulation
0
0.26 0.52 0.77 1.03 1.29 1.55 1.80 2.06 2.32 2.58 2.84 3.09
Traffic Intensity ȡ=ƣȜ/cȝ
Fig. 4. Performance comparison between theoretical and simulation CBP and BU results, where
c = 128; n = 32, and ƣ = 16.5
Table 1. Performance comparison between theoretical (theo.) and simulation (simu.) results by
a geometric group-size distribution, where c = 32; n = 16, and ƣ = 4.83682
CBP BU ASL
theo. simu. theo. theo. simu. theo. simu.
ρ (Ω ) ( Ψ1 ) ( Ψ2 ) ( Ls )
0.45 0.0374 0.0516 0.4175 0.4175 0.3839 13.4 12.2
0.76 0.1118 0.1284 0.5902 0.5902 0.5495 18.9 17.5
1.06 0.1892 0.2050 0.6889 0.6889 0.6538 22.0 20.9
1.66 0.3121 0.3277 0.7872 0.7872 0.7674 25.2 24.5
1.96 0.3589 0.3740 0.8147 0.8147 0.7996 26.1 25.5
2.57 0.4328 0.4465 0.8509 0.8509 0.8424 27.2 26.9
2.87 0.4626 0.4754 0.8635 0.8635 0.8572 27.6 27.3
3.48 0.5120 0.5225 0.8826 0.8826 0.8791 28.2 28.0
3.63 0.5226 0.5324 0.8865 0.8865 0.8833 28.4 28.2
as dotted lines with cross signs and plus signs. As shown in Fig. 4, both the CBP and
the BU rise with increasing ρ because of the limited capacity, making the simulation
results close to the theoretical results.
In the second simulation, given a carrier with a 5MHz bandwidth, the system has
the peak throughput 20.6Mbps and divides the carrier to 32 subcarriers. Then each
subcarrier supports the average data rate 659.2kbps. Assume that the arriving group
size has a geometric distribution, where c = 32; n = 16; ƣ = 4.83682019, and {xk. | 1 ≤
k ≤ 16} = {0.18899, 0.15475, 0.12663, 0.10372, 0.08486, 0.06949, 0.05698, 0.04659,
0.03812, 0.03127, 0.02558, 0.02094, 0.01713, 0.01404, 0.01149, 0.00941}. Table 1
lists some of the simulation and theoretical results, where μ = 0.00625 and λ ranges
from 0.01875 to 0.15. The table also compares other theoretical measures with those
Efficient Performance Management of Subcarrier-Allocation Systems 419
0.9
0.8
CBP
BU
0.7
0.6
CBP/BU
0.5
0.4
0.3
0.2
0.1
0
32
56
80
104
128
152
176
200
224
248
272
296
320
344
368
392
416
440
Number of subcarriers (c)
Fig. 5. CBP and BU in the OFDMA subcarrier-allocation system when the number of
subcarriers c increases
Table 2. The optimal values of c for maximizing BU with CBP ≤ 4% (CBPthreshold = 4%)
6 Conclusion
This paper presented the batch-arrival queueing model MX/M/c/c for evaluating the
performance of the OFDMA subcarrier-allocation system. The simulation results
agree with the predictions derived from the theoretical models. Additionally, the
performance measures were applied to solve the subcarrier utilization-optimization
problem that uses a given traffic load and a specified maximum CBP constraint, and
searches for the optimal number of subcarriers to maximize the BU. As a result, the
operators may utilize the results to manage their frequencies or subcarriers flexibly
and to increase their profit.
Efficient Performance Management of Subcarrier-Allocation Systems 421
References
1. Sampath, H., Talwar, S., Tellado, J., Erceg, V., Paulraj, A.: A Fourth-generation MIMO-
OFDM Broadband Wireless System: Design, Performance, and Field Trial Results. IEEE
Commun. Mag., Vol. 40, No. 9 (2002) 143–149
2. Intel Literature Center: Orthogonal Frequency Division Multiplexing.
http://www.intel.com/ netcomms/technologies/wimax/303787.pdf (2005)
3. Jamalipour, A., Wada, T., Yamazato, T.: A Tutorial on Multiple Access Technologies for
Beyond 3G Mobile Networks. IEEE Commun. Mag., Vol. 43, No. 2 (2005) 110–117
4. Dubuc, C., Starks, D., Creasy, T., Hou, Y.: A MIMO-OFDM Prototype for Next-
generation Wireless WANs. IEEE Commun. Mag., Vol. 42, No. 12 (2004) 82–87
5. Zhang, Y.J., Letaief, K.B.: An Efficient Resource-allocation Scheme for Spatial Multiuser
Access in MIMO/OFDM Systems. IEEE Trans. on Commun., Vol. 53, No. 1 (2005)
107–116
6. Song, P., Cai, L.: Multi-user Subcarrier Allocation with Minimum Rate Requests for
Downlink OFDM Packet Transmission. Proc. of IEEE VTC’04, Vol. 4 (2004) 1920–1924
7. Bakhtiari, E., Khalaj, B.H.: A New Joint Power and Subcarrier Allocation Scheme for
Multiuser OFDM Systems. Proc. of IEEE PIMRC’03, Vol. 2 (2004) 1959–1963
8. Liang, X., Zhu, J.: An Adaptive Subcarrier Allocation Algorithm for Multiuser OFDM
System. Proc. of IEEE VTC’03, Vol. 3 (2003) 1502–1506
9. Kulkarni, G., Adlakha, S., Srivastava, M.: Subcarrier Allocation and Bit Loading
Algorithms for OFDMA-based Wireless Networks. IEEE Trans. on Mobile Computing,
Vol. 4, No. 6 (2005) 652–662
10. Han, Z., Ji, Z., Liu, K.J.R.: Fair Multiuser Channel Allocation for OFDMA Networks
using Nash Bargaining Solutions and Coalitions. IEEE Trans. on Commun., Vol. 53, No. 8
(2005) 1366–1376
11. Hoymann, C.: Analysis and Performance Evaluation of the OFDM-based Metropolitan
Area Network IEEE 802.16. Computer Networks, Vol. 49, No. 3 (2005) 341–363
12. Gross, J., Geerdes, H.-F., Karl, H., Wolisz, A.: Performance Analysis of Dynamic
OFDMA Systems with Inband Signaling. IEEE Journal on Selected Areas in Commun.,
Vol. 24, No. 3 (2006) 427–436
13. Du, Z., Cheng, J., Beaulieu, N.C.: Accurate Error-rate Performance Analysis of OFDM on
Frequency-selective Nakagami-m Fading Channels. IEEE Trans. on Commun., Vol. 54,
No. 2 (2006) 319–328
14. Canpolat, B., Tanik, Y.: Performance Analysis of Adaptive Loading OFDM under
Rayleigh Fading. IEEE Trans. on Veh. Tech., Vol. 53, No. 4 (2004) 1105–1115
15. Lee, W.C.Y.: CS-OFDMA: A New Wireless CDD Physical Layer Scheme. IEEE
Commun. Mag., Vol. 43, No. 2 (2005) 74–79
Convergence Services Through NGN-CTE on the
Multiple Service Provider Environments in NGN
1 Introduction
Concepts of Convergence Service have been regarded as one of most important issues
for NGN deployment. The standardization activities of ITU-T have mainly focused on
the requirements and frameworks around the convergence services [1]. The scenario
based approach in ITU-T has accelerated their concentrations on this issue. Even with
these efforts, the actual deployment of convergence services look too far to be
realized.
On the other hand, there has been a trend of terminal devices with various
capabilities and interfaces. This trend seems to choke the end users’ desires and
preference on these various services and vice versa. However, terminals with
functions satisfying this trend might require larger capacities and accordingly very
high costs, which lead to make the deployment in the practical market even harder. A
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 422 – 430, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Convergence Services Through NGN-CTE on the Multiple Service 423
possible solution on this problem is to allot the functions between terminal side and
network side. This way also provides a breakthrough to the standardization efforts of
ITU-T on the Convergence Services in that the allotted functions in the network side
can be provided using the standardized Convergence Service scenarios.
The allotment of the service functions mainly occurs in the NGN-CTE (NGN
convergence service terminal equipment) and NGN Service Stratum. The NGN-CTE
will play an important role to consolidate the deployment of NGN at the initial phase.
At the initial stage, convergence may be initiated at the terminal side with less burden
of the network or service provider side. Thus the deployment model for NGN
convergence service terminals will be explored in this paper. Next, the way to
compose a Convergence Service using this deployment model will be explained using
the form of scenario.
Users’ Benefits from the Service Profile Mgmt. in the network side are as follows:
On the other hand, Providers’ Benefits from the Service Profile Mgmt. in the
network side are as follows:
3.1 Balance Between User and Network Sides for Convergence Service
Provision
Convergence services in NGN can be provided without changes in the network side if
multi-functional terminal can support all kinds of convergence services end users
demands. However, this approach is not regarded as practical on the cost and business
aspect of the convergence services. On the contrary, the convergence services can be
actuated using the capabilities inside NGN Service Stratum while using the low-cost
terminals at the user side.
These aspects require a kind of guideline to both the user side and the network
side. The former can be specifically embodied as the NGN-CTE while the latter as the
NGN supporting multiple NGN-SPs (NGN Service Providers).
vGSG}SGtS
z SGh
unuGG
hhhGp
Platform / Middleware
qh}hGl
ljtOkGl P
Oq}tSGqupSGtGjP
lttOw Gl P yTGvz
kGk
Device
kOkSGkP
unuGj{l
oV~Ojw|SGmSGyhtP
Figure 1 shows the logical structure of the NGN-CTE. This structure accords to the
ISO/OSI layering convention.
Convergence Services Through NGN-CTE on the Multiple Service 425
Using the deployment model shown in Figure 2, an end user service can be composed
according to the procedure as shown in Figure 3. The detailed procedure is as follows:
C andidate Servic e
Collec tion C omponent lists (optional)
Considerations on the Service User Profile are essential for the Convergence Services
to be successfully deployed in NGN. Continuous updates of the Service User Profile
are essential for the successful composition of the Convergence Services. To update
the profile effectively, the network-friendly repository tool, e.g. LDAP (Light -weight
Directory Access Protocol), is more appropriate. The update of the Service User
Profile can be actualized based on LDAP as shown in Figure 4. The LDAP server
dealing with the Service User Profile maintains the Service User Profile Information
NGN-SP (C)
Base based on the input from the participating NGN-SPs. Participating NGN-SP can
be a station of LANs or other types of networks while each NGN-SP has its own IP
address. Table 1 shows an example of the Service User Profile shown in Fig. 4.
User 1 User2 ….
Name James Smith David Miller ….
Email
[email protected] [email protected] ….
address
User Phone
010-234-5678 023-342-1090 ….
Profile number
Other
: :
attributes ….
: :
:
Connected IP 203.146.234.5 IP 127.202.189.235
User ….
Address MAC 5A:2B:23:0D:1A:2A MAC 61:2C:67:95:3A:1F
Participating IP 201.35.23.52 IP 201.35.23.52
….
SP Address MAC 1D:1B:43:25:3C:21 MAC 70:2B:1C:7F:E0:37
Participating IP 215.129.230.36 IP 131.32.21.49
….
Previously SP Address MAC NA MAC NA
Used : : : :
Service Service
Profile 1 Voice+Video Voice+Data ….
Type
Service
2006:02:23:02:27 2006:02:19:13:11 ….
Start Time
Service End
2006:02:23:02:29 2006:02:19:13:14 ….
Time
: : : ….
: : : ….
Connected IP 201.132.3.12 IP 202.121.34.2
User ….
MAC 4D:23:12:3F:2A:7D MAC 3E:F4:12:35:6B:F6
Address
Participating IP 201.35.23.52 IP 203.125.23.78
….
SP Address MAC NA MAC 5D:1B:43:25:3C:69
Participating IP 203.32.56.21 IP 201.34.28.121
Previously ….
SP Address MAC NA MAC NA
Used
Service : : : :
Profile 2 Service
Data+Video Voice+Data+Video ….
Type
Service
2006:02:22:22:31 2006:02:01:16:12 ….
Start Time
Service End
2006:02:22:23:39 2006:02:01:17:02 ….
Time
: : : ….
: : : : :
the proposed deployment model seeks the appropriate NGN-SPs among candidates
after receiving Convergence Service Request from the end user. The list of NGN-SP
must be maintained for this search process. Pre-registration of NGN-SPs into the list
is also needed.
NGN-CTE can request the convergence services customized to the end users
preference. The preference of end users includes the selection of NGN-SPs that
provide same services at the lower cost than the competitive ones. Customization and
Personalization of the service provision can be easily approached using a kind of GUI
at the NGN-CTE. The detailed aspect of the GUI is for further study.
6 Conclusion
A deployment model and a service composition scenario are proposed for providing
the Convergence Services to the end users. The scenario shows that the proposed
deployment model can be useful for creating new Convergence Services over NGN.
In addition, the LDAP-based procedure for updates of Service User Profile is
presented. Customization and Personalization of service provision is expected to be
studied more for completeness of this approach. More study is also needed to clarify
the details to this brief outline of the Convergence Services in NGN.
Acknowledgments. This work was supported by the Inje Research and Scholarship
Foundation in 2004.
References
1. ITU-T FGNGN-OD-00248R1 (2005), Revised working draft TR-CSF Version 3.0 (NGN
Release 2)
2. ITU-T Recommendation Y.2011 (2004), General principles and general reference Model for
NGNs
3. ITU-T FGNGN-OD-00223, Draft FGNGN-FRA Version 6.3.
Proposal of Operation Method for Application
Servers on NGN Using Unified Management
Environment
1 Introduction
1.1 Next Generation Network
Recently, a lot of network carriers such as British Telecom (BT) and Nippon
Telegraph and Telephone Corporation (NTT) have been aiming at the introduc-
tion of NGN, which resolves the problems of existing public switched telephone
networks (PSTN) and enables various services such as multicasting multime-
dia content between interest-sharing communities over internet protocol (IP)
based networks. To achieve the NGN successfully, such carriers are discussing
the standardization of NGN architecture at conferences such as International
Telecommunication Union Telecommunication Standardization sector (ITU-T),
European Telecommunications Standards Institute (ETSI), and Telecom and
Internet-converged Services and Protocols for Advanced Networks (TISPAN)
[1][2][3]. Researchers at TISPAN define the NGN architecture (Fig. 1) as a lay-
ered structure with an IP-based transport layer and a service layer.
In detail, they define IP-connectivity assurance systems in the transport layer
and service provision components in the service layer. The group of service pro-
vision components for business and end users is in the service layer of an NGN-
based infrastructure. For example, Core IP Multimadia System (Core IMS) and
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 431–440, 2006.
c Springer-Verlag Berlin Heidelberg 2006
432 A. Yoshida, Y. Miyoshi, and Y. Otsuka
The NGN has a complex architecture to provide high service provision perfor-
mance. Services on the NGN are provided by combining several service provision
components that work on NGN servers. For providing seamless services to NGN
users, network carriers should manage each component properly; thus, the op-
eration and management tasks are difficult. This complexity causes high NGN
maintenance costs for network carriers. We need to reduce the cost of achieving
the NGN.
Researchers on TeleManagement Forum (TMF) standardize network opera-
tion processes to reduce the cost of operating the NGN. This standard is NGOSS
[4]. NGOSS has four frameworks.
(Fig. 2) [5]. By using eTOM, we can map between business processes and oper-
ations support system (OSS) functions.
Symbolized by frameworks such as eTOM, operation software tools are clas-
sified by operation functions such as those of the service management layer and
the resource management layer. Operators in network carriers need to use dif-
ferent tools in complex procedures if the management objects are in different
layers, even when the same services are operated. To integrate operational tasks
is difficult for operators. For example, they need to use IP reachability check
tools such as ping and TCP/UDP state check tools such as netstat for managing
a server.
For resolving the above problems, we propose the examination of the unified
operation method, which examines service provision processes of servers on the
NGN service layer. Our purpose is to improve operator usability. We propose a
unified operation method for application servers on NGN.
In section 2, we introduce design of an automatic management technique for
server components such as hardware, OS, and applications. The design of the
proposed method is described in section 3. We outline our verification of our
proposal in section 4. After that, we examine the achievability of the proposed
method in section 5 and show the experimentl results in section 6. And we
discuss the results and achievability of our proposal in section 7. In section 8,
we conclude our work.
2 Related Work
To reduce server management costs, there is design of an automatic management
technique for server components such as hardware, OS, and applications. This
technique is called ISAC [6], and it provides automatic server management by a
PCI-X card-based system that plugs into an available slot on the server. Accord-
ing to user-defined management scenarios, a PCI-X card and software installed
434 A. Yoshida, Y. Miyoshi, and Y. Otsuka
on the server manage the server components and communicate with each other
to enable automatic management. A comparison between the proposed method
and ISAC is shown in Table 1. A purpose of ISAC is an independence of man-
agement method from a server condition. It doesn’t consider operator usability
using the unified environment.
method uses the command line interface (CLI) based supprt technique that is
used in operations such as the configuration or validation of NEs.
This enables operators to execute service operations using the same manipu-
lation as that for NE operations and to unify the operation environment. This
unified operation environment achieves IP network management in servers and
control by a single software tool. In addition, this control brings about the follow-
ing effects. For service developers, promoting efficiency such as that achieved by
automation becomes easy. Moreover, the scope of the automation is expanded.
For operators, there is no longer a need to develop and manipulate multiple tools
and specific clients. As a result, NGN operations become efficient.
has executed in a script format, replays the recorded process, and executes that
process on a schedule determined by a timer. Therefore, once a manipulation is
performed on the CLI, operators can automatically execute that manipulation
that has already been performed. This significantly reduces the operation load of
routine work. In addition, our technique focuses on the script format. By defin-
ing the script in a operator-friendly format, flexibly modifying the commands,
the conditions, and the parameters recorded in the script becomes easy. This
enables operators to conduct tests while reviewing and modifying the test en-
tries on site. In particular, during the validation, the time lag caused by software
programming development is reduced.
4 Outline of Verification
We executed an experiment to verify the realizability of the proposal method.
We chose components of web services provided on the application server, as the
subjects of this verification.
In general, a web service consists of application servers that provide a web
service, a web site that provides services composed of web services, and a client
that uses web services.
An application server that provides web services offers a function to send
WSDL messages and a function to send and receive SOAP messages to provide
web services. To verify the application server, monitoring and controlling these
two functions are required. Web services use several resources, so when a failure
of a web service occurs, we need to check the statuses of the resources used by
this service and specify the causes of the failure. The statuses of these items
need to be checked to verify the application server.
1. Reachability of IP packet
2. HTTP daemons
3. Platforms used to provide Web Service (e.g., tomcat)
Proposal of Operation Method for Application Servers on NGN 437
4. WSDLs
5. Web services
We need to use a different tool to check each item shown above to verify
the application server by existing verification methods. Some points need to be
checked manually, so we think there are some processes that need to be changed
to automatic execution.
5 Experiment
Here, we explain the procedure of verifying an application server that provides web
services. First, we check the reachability of an IP packet by sending a ping com-
mand to the application server. Next, we check whether HTTP daemons on the
application server are running by sending specified HTTP requests to the applica-
tion server, which platform is being used to provide a web service on the applica-
tion server, and whether WSDLs, which describe web services on the application
server, can be obtained. We check whether the syntax of the obtained WSDLs is
correct. Therefore, if the syntax is correct, we can extract the operation methods
from each WSDL. We execute operations in a particular WSDL in the correct order,
and check each response of the operations. Once the obtained WSDLs are checked
appropriately by the above process, we store the XML messages used by the check
(the obtained WSDLs, request messages, and response messages) as script files. By
partially rewriting these scripts and resending them, we can automatically check a
particular application server. In this paper, we experimented with an application
server to verify the procedure described above.
In the experiment, we use telnet software to obtain WSDL text and to request
web service operations. To check the function to obtain WSDL texts, we capture
HTTP requests generated by a web browser, and we resend these requests using
telnet software. We verify the realizability of checking the function to obtain
WSDL texts (Fig. 5) using this operation.
We capture SOAP messages generated by dedicated client software to check
the function to execute web services, and we resend these messages using telnet
software. By this operation, we verify the realizability of checking the function
that executes web services (Fig. 6).
In these experiments on web service and web service tests, we can achieve text
processing of messages by partially rewriting SOAP messages, which is carried
out by partially changing the SOAP message parameters. We verified these two
application server functions, to supply WSDL and to execute web services, with
the following check items. Thereby, the availability of the proposed method was
estimated.
– check item 1: Can we monitor these functions? (Monitoring)
– check item 2: Can we experiment various tests? (Experiment)
– check item 3: Can we execute the automation of these two functions? (Au-
tomation)
438 A. Yoshida, Y. Miyoshi, and Y. Otsuka
sider whether operators can execute text processing of request messages, and
whether the experiment can be executed automatically.
Automation: Estimate the possibility of automating the above monitoring and
experimental functions from the results of the experiment. In addition, we con-
sider requirements for the automation.
6 Experimental Results
Results of the experiment are shown in Table 2. In the first step, we performed a
WSDL function test. We sent a sequence of captured HTTP requests that were
generated by a browser software to call an application server. In this test, we
confirmed WSDL functions on application servers.
In the second step, we checked the service provision capability using a test
tool. We used a dedicated web service client software and captured transactions
of SOAP messages between this client and the server. Next, we sent captured
request messages to the server, and captured correct responses from the server. In
this trial, we confirmed that there was no need for setting environment variables
or compiling client software compared with using a Java client. Therefore, the
capability of constructing a more generic testbed for the application server was
demonstrated. In addition, many packets were generated in one request, but
there was no effect on the response of the server. We edited various SOAP
message parameters contained in XML tag containers and HTTP headers. Then
we confirmed that we could use a modified request.
7 Discussion
Periodic WSDL monitoring is useful because HTTP requests are normally used.
Resending a request with a different parameter is not required, so WSDL is not
changed until the alteration of service specification. We consider two WSDL test
capabilities can be automated by script using specified URL of WSDL. In this
way, we can use retrieved WSDL for remote monitoring functions and automated
checking of functions of web services in test scripts.
On the other hand, we can make a function for monitoring web services using
captured requests and responses by normal client and the server. Increasing
the efficiency of building a set of correct SOAP messages for beginning the
test is important. In the experiment on service level transactions, we confirmed
that rewriting HTTP headers and parameters is easy; therefore, testing web
440 A. Yoshida, Y. Miyoshi, and Y. Otsuka
service messages using automatic parameter rewriting with text processing tools
is feasible. HTTP header translation is simple and a parameter type is required
in the SOAP message. In these web service test functions, the operator must
define the range of parameters before the test, and the software testing tool will
make various parameters automatically. We believe that monitoring and testing
web services using the same method as that of monitoring of network elements
is achievable (Table 3).
8 Conclusion
We defined a method of verifying web services. With this method, we were able
to reduce the work and amount of time to perform tests of web services arranged
in an IP network. Test functions improved web service tests as listed below.
– This method makes a unified operating environment for communication func-
tions and service functions using text processing.
– Automatic operation is easily performed.
– There is no need to develop dedicated software tools.
In conclusion, we consider that we have achieved a test function with auto-
matic modification of parameters.
References
1. A. Kurokawa and I. Azuma: The NGN standardization trend in ETSI TISPAN (In
Japanese) NTT Technical Jornal Vol. 18, No. 4, (2006)
2. ETSI ES 282 007: Telecommunications and Internet Converged Services and Pro-
tocols for Advanced Networks (TISPAN) IP Multimedia Subsystem Functional Ar-
chitecture. (2005)
3. ITU-T Y.2001: General overview of NGN.
4. TM Forum: New Generation Operational Support System (NGOSS) PART 1 (Busi-
ness Case), TMF051, (2001)
5. TM Forum: Enhanced Telecom Operations Map (eTOM) The Business Process
Framework. Addendum D, Process Decompositions and Descriptions Release 5.0,
GB921D, (2005)
6. T. White, D. Carvert, and J. Litkey: Design of an Autonomic Element for Server
Management. Proceedings of the Second International Conference on Autonomic
Computing (ICAC’05) (2005)
7. Y. Miyoshi and T. Kimura: Interface Blending/Diagnosis Technology for Automat-
ically Generating an Interface Adapter. NTT Technical Review. Vol. 3. No. 10.
(2005)
IP/WDM Optical Network Testbed: Design and
Implementation
1 Introduction
The ever increasing demand for bandwidth, flexibility and reliability has led to the
proposal and implementation of new network architectures. Due to its potential
advantages, optical networks have been considered in situations where there is need
of high bandwidth and low latency communication. Among many proposals, IP-
centric control plane within WDM optical networks is a widely considered solution
for dynamic provisioning/restoration of lightpaths [1].
Several initiatives to build WDM backbones are under development [2-3]. Passive
Optical Network (PON), next generation SONET/SDH generic framing procedure
(GFP) and Ethernet Resilient Packet Ring (RPR) are helping further streamline
vertical "data optical integration". Given this evolution, there is a growing need for
testbed development in order to provide much-needed "proved-in" value [4].
The physical layer of optical networks is composed of photonic devices, which are
rather expensive. Thus, the high costs of the implementation of this kind of network
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 441 – 450, 2006.
© Springer-Verlag Berlin Heidelberg 2006
442 H.A.F. Crispim et al.
difficult the research in this field. A possible way to overcome this problem is the
development of simulation scenarios, tools and techniques. This simulations,
however, requires a careful design and study as to achieve a realistic scenario, where
details of the planes and their elements are well represented.
In this paper we describe the design and implementation of a transparent IP/WDM
Optical Network Testbed. The implemented testbed (designated hereon as LabCom
testbed) considers the control plane and the management plane in a centralized model.
For the management plane the development of a friendly and intuitive web-based O-
UNI (Optical – User Network Interface) is presented. Our system provides the
network manager with management and security mechanisms. The network is
presented in figure 1.
2 Network’s Architecture
We show our built scenario in figure 2. In this proposed architecture, the transport
network is formed by five nodes. Each node is connected in one Ethernet switch by a
VLAN connection, forming five VLAN connections. The control plane and
management functions are implemented in a personal computer (PC), forming a
centralized network architecture.
The logical topology is established by VLANs. Each VLAN corresponds to an
optical node link (OXC) of the transport plane. The sixth link connects the transport
plane to the control plane. This ensures the isolation of the broadcast domains and
also, it limits the transmission rate for each VLAN. We realized tests with ports
working at a rate of 1 Mbps.
Each optical node is simulated by a IBM PC PIII with 128 MB of RAM, processor
of 700 MHz and a 10/100 Mbps Ethernet network card connected to a switch. The
IP/WDM Optical Network Testbed: Design and Implementation 443
computer which executes the control and management system is a HP PIV 1.8 GHz
with 256 MB of RAM and a 10/100 Mbps Ethernet network card connected to a
switch.
The management plane works based on fail management. The management plane
consists of the network configuration program, the performance monitoring, the log of
changes and the access security system. The System Management Application Entity
(SMAE) is separated in two parts. Those parts are implemented by different processes
(in the application layer), situated at the management unit and at the agent unit [5].
Thus, we have a centralized management process. The management objects (OXCs.
Amplifiers and Transponders) register their parameters in their own information
management base that will be distributed among the managing entity and the agent in
each optical element. The information which corresponds to the physical elements
read status is transmitted in the XML format.
Optical User Network Interface (O-UNI Web). The routes can be requested
graphically by means of the UNI-Web interface. Therefore, the computer hosting the
management functionalities in our LabCom Testbed can be accessed by HTTP. The
proposed UNI-Web based interface [6] facilitates the interaction of the user with the
transport network, allowing an efficient and friendly operation. Another functionality
of our graphical interface is to make the auditing process completely transparent to
the network manager.
UNI-Web can create or eliminate optical links, get information on the network and
nodes state, of ports and switches, as well as visualizing the occupation of the
wavelengths in each fiber. It can also provide the network manager with information
on which requests were not taken care of and other administrative information. He
can also audit the users and executed commands, among others functionalities.
Figure 3 shows the screen of the UNI-Web.
444 H.A.F. Crispim et al.
The graphic environment was developed in web model using Java Server Pages –
JSP technology. In this interface, the user is authenticated by means of a password.
The system monitors users actions through auditing mechanisms made in a relational
model and implemented in PostgreSQL V8.1.2.
The control plane, as explained before, was implemented in a centralized fashion. Its
main function is to reroute failed connections in the event of data plane failures [7].
We defined a UNI interface as an overlay model. The interaction between the
transport and control plane follows the client/server model.
For security reasons and reliability, the control system permits only one TCP/IP
connection with the management system.
The communication between the control system and the agents implemented in the
optical devices occurs over UDP/IP connections, since in our application the error
rates will be negligible. This system is implemented based in three threads, each one
with the following functions: a) to connect to the management system and implement
the business rules; b) to manage the network link stability by means of implementing
a Hello protocol; and c) to communicate with various nodes compounding the
physical network as to manage actions such as sets, gets e traps operations.
The controlling is performed in the following logical order: a) initially the
network’s file configuration is read, it describes the physical topology and specifies
the physical links between the optical devices; b) the file topology content is parsed,
in order to evidence errors; c) the shortest paths between network elements are
computed by using Dijkstra algorithm; d) alternative and disjoint paths are computed.
Those paths are then made ready to substitute the original paths in case of link faults;
e) the controlling mechanism (working in standby) is then ready to receive the
lightpath allocation and deletion requests.
The whole control system was developed in C++ KDE environment on Linux
system operational - Fedora.
IP/WDM Optical Network Testbed: Design and Implementation 445
The physical layer elements were simulated using a graphic environment in Java. We
implemented three devices: the optical cross-connects (OXCs), the amplifiers and the
transponders. Each optical node is composed by: an OXC capably to commute eight
lambdas from ITU grid, eight transponders to adapt the signal wavelength to the ones
used in the network and eight amplifiers. Specific optical nodes can be simpler than
the structure we just described, i.e. it need not contain a transponder.
The optical components are object-oriented modeled, with classes that implement
the functions of each element of the optical node. A graphic application of the model
was developed in a way that allows a faithful reproduction of the functions of the
optical elements and the visualization of system actions that follows.
The Figure 4a shows an optical node with the received lightpaths configuration
(RX), from the control plane, and the operation status (TX). Other actions, such as
fault simulations and configuration parameters of the amplifiers, transponders and
OXC can be simulated. The figure 4b presents a simulated failure.
Our controls system is composed of three basic protocols. One implements and
creates optical pathways, the second manages the links, and the third activates an
alternative route in case of fault (protection route). Figure 5 represents the basic
structure of the physical site simulation.
Each physical element can simulate the following functions: on/off, optical ports
read, alarm to power fault at in/out optical;
446 H.A.F. Crispim et al.
Faults are introduced in the system by using a graphical interface which activates
the control system by sending a trap message. Specific actions are then launched by
the control system.
We used the Linux Fedora Release 2.0 distribution as our operational system,. The
relational data base used was the PostgreSQL V8.1.2, and Apache Tomcat 4.1 was
also used as our web server.
We used the environment KDE 3.5.2 with compiler GCC 3.4.6 and for the
development of the functions of high performance. For Java programming it was used
the eclipse-SDK-3.1.2-linux and J2SE 1.4.2_10 SDK environment.
To collect the network’s information, monitoring of various package and filters,
related with the protocol IP, it was used the Ethereal version 0.9.14 environment.
3 Testbed Working
The topology discovery is realized in a hybrid form. The management system sends a
message Hello for its managed network nodes (the IP addresses of those nodes were
obtained by using DHCP) so as to discover the active elements. The management
system repeatedly sends hello messages each 30 (thirty) seconds to confirm the link
stability of each element in the network.
In case three consecutive hello messages are not acknowledged by a node, a fault is
assigned to it. This event is informed to the manager by a graphic interface.
Besides the detections of actives elements, the system has a configuration file,
produced by the network’s manager, which has the whole physical topology. When
the activity of each optical element is detected, the system is assured that the network
has no fault and that ligthpaths can be securely implemented.
To create a lightpath, a client, through the web interfaces, make a lightpath creation
request to the management system. The management system analyzes the syntax of
the solicitation and activates the control system, if the request is correct. Otherwise, it
answers to the client informing a problem with the solicitation.
After a lightpath creating request, the control system uses the Routing and
Wavelength Assignment - RWA algorithmi, based in Dijsktra, to create a lightpath
using the First-Fit strategy to allocate a wavelength.
After it has the complete knowledge about the lightpath, the control system sends a
message to each node, which belongs to the path, telling the node to switch its optical
key with its respective wavelength. Then after the switching is proceeded the node
responds to the control system. The control system consider the path as operational
when it receives a right answer from all nodes in the path.
IP/WDM Optical Network Testbed: Design and Implementation 447
After that, the control system sends a message SETPORT to each node, which will
answer with a SETPORTOK or a SETPORNOK message.
The message SETPORTOK sets a lightpath as active, to the control system. A
SETPORTNOK answer makes the control system reject the realized request and make
all keys to turn to the original states.
For this operation, the management system asks the lightpath deletion to the control
system, which, by its turn, identifies the requested path, its wavelength and respective
nodes. So, the system control tells each optical cross connect to switch the optical
keys for the original states by sending an UNSETPORT message.
After each switching, the node responds to the system control. The system control
marks the lightpaths as deleted when all nodes respond with an UNSETPORTOK
message.
The UNSETPORTOK message corresponds to turning the lightpath inactive and
the UNSETPORTNOK message makes the control system disrespect the performed
request and reset the keys to the original states.
Each activity element (OXC, amplifier or transponder) can perform fault simulations,
by using traps. The following problems can be simulated:
1. on/off;
2. read errors on optical ports;
3. power fault on in/out optical ports;
4 Tests Results
We performed our tests with the help of a program which belongs to the management
system. This program executes a random choice of the source and target nodes. Then,
the system controls make the best choice for the lightpath (path and lambda) to be
used.
On the tests, we made thirty random requests and the network saturated, on
average, with twenty eight requests.
448 H.A.F. Crispim et al.
The Table 1 presents the results of the requests made to the management system.
On the first column we have the number of request, on the second column the time in
milliseconds to attend de request, on the third the number of nodes that belong to the
lightpath and, in the last column, the complete lightpath (with node number, lambda,
in/out optical port on each OXC).
In this specific test, with thirty random requests, twenty six were attended and four
were denied due to network saturation. Then, we observe the need of an intelligent
system for wavelength allocation. However, this problem is beyond the scope of this
work.
The mean time for attending one request was 321 ms and the average number of
hops was three.
The Table 2 presents the all packets used for attending the requests (twenty six).
On this data we can view that the protocol consumed thirty percent of the whole
traffic on the network.
IP/WDM Optical Network Testbed: Design and Implementation 449
Lightpaths creation
500
426 435 448 434
400 370 476
364 351
319 304
Time - ms
0
1 3 5 7 9 11 13 15 17 19 21 23 25
Request number
5 Conclusions
simple modeling using XML for the exchange of information. Because we used the
UDP protocol, our solution possesses high performance and an intelligent provision
of optical resources.
References
1. B. Rajagopalan, D. Pendarakis, R.S. Ramamoorthy, D. Saha, K. Bala, “IP over optical
networks: Architectural aspects” IEEE Communications Magazine, no.9, pp. 94-102,
(2000).
2. D. Cavendish, “Evolution of optical transport technologies: from SONET/SDH to WDM”
IEEE Communications Magazine, no.6, pp.164-172, (2000).
3. R. E. Wagner et al., “MONET: Multiwavelength optical networking” IEEE JSAC, vol. 14,
pp. 1349-1355, (1996).
4. Optical networking testbeds: experiences, challenges, and future directions (Part 1), Mohan,
G.; Ghani, N.; Saradhi, C.V.; Stavdas, A. IEEE Optical Communications –Page(s): 52- 53
(2005)
5. W. Stallings, SNMP, SNMPv2, SNMPv3, and RMON 1 and 2. Reading, MA: Addison-
Wesley, (1996).
6. H.A.F. Crispim, Eduardo T. L. Pastor, H. Abdalla Jr, and A.J.M. Soares "Interface de
Usuário para rede óptica IP/WDM em ambiente Web", XXII Simpósio Brasileiro de
Telecomunicações - SBrT’– Brazil (2005).
7. Guangzhi Li, Jennifer Yates, D. Wang, Charles Kalmanek, “Control Plane Design for
Reliable Optical Networks” IEEE Communications Magazine, pp. 90-96, (2002).
Voice Quality Management for IP Networks
Based on Automatic Change Detection
of Monitoring Data
1 Introduction
– End-to-end delay
is the time required for a voice packet sent by the caller to reach the callee.
(This is equivalent to the difference between the packet arrival time at the
callee and the packet timestamp that the caller puts on the transmission.)
Real-time communication quality falls to a critical level when the delay ex-
ceeds a specific value (e.g. 150msec).
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 451–460, 2006.
Springer-Verlag Berlin Heidelberg 2006
452 S. Imai et al.
– Jitter
is the variation in packet arrival times. (This is equivalent to the difference
between each end-to-end delay and the average end-to-end delay.)
If the packet arrives too early or too late, the playout quality will be bad,
because voice codecs require a steady packet stream to provide adequate
playout quality.
– Packet Loss
is a phenomenon in which voice packets are discarded on the IP network.
Voice quality deteriorates because jumpiness and noise are caused by losing
a part of the voice signals.
The behavior of these packet parameters depends on changes in the network
conditions. In order to sustain VoIP quality, it is necessary to detect the change
in network conditions before the VoIP quality deteriorates, and prevent any
deterioration in quality.
Our study assumes an active measurement scheme by sending test packets
as VoIP quality monitoring schemes, and focuses on a change detection scheme
for network conditions (e.g. traffic load) based on analyzing the behavior of
the monitoring data. In a change detection scheme, we should preset a suitable
threshold to detect changes in the behavior of the monitoring data associated
with a change in network conditions. However, as IP networks become large in
scale and application traffic becomes more diverse, the behavior of the moni-
toring data (e.g. end-to-end delay and jitter) detected by active measurement
will become nonidentical and more complex. Therefore, it is not easy to preset
proper thresholds for data monitoring at all measurement points.
In this paper, we focus on the fact that jitter is extremely sensitive to changes
in the network load, and propose a VoIP quality management scheme that detects
the change in the behavior of the jitter associated with a change in the network
load automatically and in real time by applying a statistical scheme.
The contents of this paper are organized as follows. Section 2 gives the VoIP
quality management model and technical issues. Section 3 describes our scheme
based on statistical tests. Section 4 presents some simulation results, and Section
5 concludes the paper with a summary and future studies.
– The management system or the network operators control the specific target
(e.g. SIP server) based on the trigger of the change alarm.
In this section, we describe our proposed scheme. Our scheme can detect auto-
matically and instantaneously changes in the behavior of the jitter measured by
the monitoring agent. The algorithm is performed as follows. (Fig. 2)
(i) Measure the end-to-end delay x[n] in time n, and calculate the jitter p[n]
from x[n] by an exponentially-weighted moving average (EWMA).
(ii) Calculate the sample variance s21 [n] of the residual errors in the AR model
using an adaptive lattice filter [4,5].
(iii) Calculate the statistical value and detect the instantaneous change (outlier)
from the basic variance s20 by the F-test [6].
(iv) Detect any stationary change in s21 [n].
(v) Update the basic variance s20 when a stationary change is detected.
The detailed algorithms on this flow are described below.
m
pt = − ai pt−i +
t . (1)
i=1
where
t is the normally-distributed residual error with a mean of 0, and variance
σ 2 . W is the sampling data size for the estimation of AR parameters ai (i =
1, · · · , m), the AR order m is a time-variable parameter which is optimally set
by Akaike’s Information Criterion (AIC) [3] in time n. In this paper, the jitter
p[n] is defined as p[n] = x[n]− T [n], where the mean T [n] of the end-to-end delay
x[n] is defined based on EWMA with the forgetting value 0 < λ << 1, which
gives
0≤m<M
γm [−1] = ρm [−1] = Δm+1 [−1] = 0
0≤n
γ0 [n] = 0,
0 [n] = ρ0 [n] = p[n]
E0 [n] = E0ρ [n] = ωE0 [n − 1] + p[n]2
0≤m<M
m [n]ρm [n−1]
Δm+1 [n] = ωΔm+1 [n − 1] + 1−γm [n−1]
km+1 [n] = ΔEm+1
[n] ρ
km+1 [n] = E Δm+1 [n]
ρ
m [n] m [n−1]
ρ
m+1 [n] =
m [n] − km+1 [n]ρm [n − 1]
ρm+1 [n] = ρm [n − 1] − km+1
[n]
m [n]
ρ
Em+1 [n] = Em [n] − km+1 [n]Δm+1 [n]
ρ
Em+1 ρ
[n] = Em [n − 1] − km+1
[n]Δm+1 [n]
ρm [n]2
γm+1 [n] = γm [n] + Emρ
[n]
The above-described parameter (1−ω)∗Em [n] with the forgetting value 0 <<
ω < 1 in the adaptive lattice filter algorithm is equivalent to the sample variance
of residual error
m [n] in the number of samples defined by W = 1/(1 − ω). In
this paper, the order m to minimize AIC values expressed by
2(m + 1)
AICm [n] = min[ log ((1 − ω) ∗ Em
[n]) + ] (3)
m 1/(1 − ω)
is defined as the optimal order m̂. In our approach, we detect changes in the
behavior of the jitter by observing this sample variance (1 − ω) ∗ Em̂
[n].
Outlier Detection. We treat the scheme which tests the statistical differences
between two sample variances (s20 and s21 ) as a framework to detect any outlier
to the sample variance (s21 ) based on the basic sample variance (s20 ).
The basic idea of outlier detection is described as follows. Based on the ev-
idence that the residual errors in the optimal AR model belong to a normal
distribution, we define basic n0 -samples of the residual errors as
[Basic Samples]
01 , · · · ,
0n0 ∼ N {0, σ02 }, i.i.d (4)
[Test Samples]
11 , · · · ,
1n1 ∼ N {0, σ12 }, i.i.d (5)
Testing the differences between the sample variances s20 from the basic samples
and the sample variances s21 from the test samples boils down to carrying out
an F-test on the following hypotheses:
H0 : σ02 = σ12
(6)
H1 : σ02 = σ12
The statistical value based on the null hypothesis (H0 ), which is expressed by
n1 s21
n1 −1
F0 = n0 s20
(7)
n0 −1
Because we premise on the basis that the both s20 and s21 are derived by the
adaptive lattice filter, the sample variances s21 are calculated in time n, and both
n0 and n1 are equivalent to 1/(1−ω). Therefore, the outlier detection rules based
on the basic variance s20 with significance level α are defined as
s21 [n] ω ω
>F α/2
, ⇒ An upper outlier has occurred (9)
s20 1−ω 1−ω
s21 [n] ω ω
< Fα/2 , ⇒ A lower outlier has occurred (10)
s20 1−ω 1−ω
Updating the Basic Variances. The update rule for the basic variance s20 is
described as below. We expand a ”scheme to detect an outlier” to a ”scheme to
detect a stationary change”. The update rule is
Voice Quality Management for IP Networks 457
Fig. 3. F-distribution
The problem of testing these hypotheses can be resolved using statistical values
which are expressed by
r1 − α/2 r2 − α/2
Z1 =
, Z2 =
(14)
α/2(1 − α/2)/N α/2(1 − α/2)/N
If the trial number N is sufficiently large, the statistical values (Z1 and Z2 )
belong to a standard normal distribution. When the upper α̃ point of a standard
normal distribution is Z α̃ , the rules to detect whether or not a stationary change
has occurred are as expressed below.
Z1 > Z α̃ ⇒ stationary change has occurred on the upper side (15)
Z2 > Z α̃ ⇒ stationary change has occurred on the lower side (16)
When a stationary change is detected by the above-described rules, we update
the basic variance s20 to
458 S. Imai et al.
– s20 ∗ F α/2 ω ω
1−ω , 1−ω ⇐ change on the upper side
– s20 ∗ Fα/2 ω ω
1−ω , 1−ω ⇐ change on the lower side
In our approach, the estimation of the both rates in Equation(11) are deter-
mined in real time. If D[n] is a discrete parameter that is 1 when an outlier is
detected, and 0 when no outlier is detected, the rate r[n] in time n is calculated
using the forgetting value 0 < η 1 as below.
r[n] = (1 − η)r[n − 1] + ηD[n] (17)
Where the trial number N is equivalent to about 1/η. In this paper, we set η as
α/2 based on the significance level for outlier detection.
Relationship Between Load and Jitter. The jitter, which is calculated from
the end-to-end delay, has the following characteristics as shown in Fig. 5.
– As the traffic load increases, the jitter also increases.
– The jitter in the mixed traffic case is larger than the jitter in the case with
only VoIP traffic.
Because our scheme detects behavioral changes in the jitter associated with
sharp changes in the traffic load by observing the behavior of the residual errors
in the AR model, we show the relationship between the traffic load and the
residual errors. The relationship between the means of the traffic load and the
variances of the residual errors per 1 [sec] are shown in Fig. 6. The correlation
coefficients below an about 90% load in Fig. 6(a) and Fig. 6(b) are 0.912 and
0.872, respectively. These results show a strong correlation between the traffic
load and the variances of the residual errors in both scenarios. Therefore, it is
effective to observe the variances of the residual errors in order to detect sharp
changes in the traffic load.
Voice Quality Management for IP Networks 459
Detection Results. The detection results for the end-to-end delay in Fig. 4
are shown in Fig. 7. The change values in these results are the sum of the change
amounts, which indicate +1 when it is detected that the jitter increased, and
−1 when it is detected that the jitter decreased.
In the simulation results, we can verify that our scheme detects that the vari-
ances of the residual errors (i.e. the behavior of the jitter) have changed after
the traffic load changes sharply, for example, at about 520 [sec] in Fig. 7(a) or
at about 550 [sec] in Fig. 7(b). Therefore, our scheme enables us to detect auto-
matically the changes in the behavior of the jitter associated with the changes in
460 S. Imai et al.
the traffic load, without presetting a threshold for the jitter, based on recursive
calculations.
5 Conclusion
In this paper, we focus on the measurement of jitter by a monitoring agent,
and propose a change detection scheme based on time-series modeling and a
statistical approach, which detects automatically and instantaneously changes
in the behavior of jitter associated with sharp changes in the traffic load. From
the simulation results, we were able to verify that our scheme is very effective as
a scalable system, because it can be operated without presetting any information
about the monitoring data.
Future studies will focus on research and development of a control scheme to
avoid any deterioration in VoIP quality based on this change detection scheme.
References
1. Lui, C.L., Fu, T.C, and Cheung, T.Y.: Agent-Based Network Intrusion Detection
System Using Data Mining Approaches. Third International Conference on Infor-
mation Technology and Applications (ICITA), Vol. 1. (2005) 131-136
2. Petrovsky, M. I.: Outlier Detection Algorithms in Data Mining Systems, Program-
ming and Computer Software, Vol. 29, Issue. 4. (2003) 228-237
3. Chatfield, C.: The Analysis of Time Series: An Introduction, Sixth Edition. Chap-
man & Hall/CRC. (2004)
4. Friedlander, B.: Lattice Filters for Adaptive Processing. Proceedings of IEEE,
Vol. 70. (1982) 829–867.
5. Singer, A.C.: Universal Linear Prediction by Model Order Weighting. IEEE Trans,
Signal Process, Vol. 24, No. 10. (1999) 2685–2699.
6. Bulmer, M.G.: Principles of Statistics. Dover Publications, Inc. (1979)
Parameter Design for Diffusion-Type
Autonomous Decentralized Flow Control
1 Introduction
The rapid spread of the Internet will necessitate the construction of higher-speed
backbone networks in the near future. In a high-speed network, it is impossible
to implement time-sensitive control based on collecting global information about
the whole network because the state of a node varies rapidly in accordance with
its processing speed although the propagation delay is constant. If we allow
sufficient time to collect network-wide information, the data so gathered is too
old to use for time-sensitive control. In this sense, each node in a high-speed
network is isolated from up-to-date information about the state of other nodes
or that of the overall network.
This paper focuses on a flow control mechanism for high-speed networks.
From the above considerations, the technique used for our flow control method
should satisfy the following requirements: (i) it must be possible to collect the
information required for the control method, and (ii) the control should take
effect immediately.
There are many other papers reporting studies on flow control optimization
in a framework of solving linear programs [1,2,3]. These studies assume the
collection of global information about the network, but it is impossible to achieve
such a centralized control mechanism in high-speed networks. In addition, solving
these optimization problems requires enough time to be available for calculation,
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 461–470, 2006.
c Springer-Verlag Berlin Heidelberg 2006
462 C. Takano, K. Sugiyama, and M. Aida
temperature distribution
iron bar
heat source
node and that in an adjacent node. Thus, the distribution of the total number
of packets in a node in the network becomes uniform over time. In this control
mechanism, the state of the whole network is controlled indirectly through the
autonomous action of each node.
Our previous studies show that our flow control mechanism with certain pa-
rameter settings works well in high-speed networks. However, to apply DFC to
actual networks, it is necessary to clarify how to design parameters in our control
mechanism. This is one of central issues to be solved for applying DFC to actual
networks. In this paper, we investigate the appropriate value of a parameter in
DFC and propose a design policy of the value.
2 Preliminary
As mentioned above, the framework of our flow control model involves both
autonomous decision-making by each node and interaction between adjacent
nodes. There is no centralized control mechanism in the network.
Next, we explain the details of DFC. The transmission rate Ji (α, t) of node i
at time t is determined by
where Li (t) denotes the value of the available bandwidth of the link from node
i to node i + 1 for target flow at time t, ni (t) denotes the number of packets
in node i at time t, ri (t − di ) is the target transmission rate specified by the
downstream node i + 1 as feedback information, and di denotes the propagation
delay between nodes i and i + 1. Determination of Li (t) is explained later. In
addition, ri (t − di ) and ni+1 (t − di ) are reported from the downstream node
i + 1 as feedback information with propagation delay di . Parameter α (≥ 1),
which is a constant, is the flow intensity multiplier. Parameter Di is chosen to
be inversely proportional to the propagation delay [6] as Di = D/di , where D
(> 0), which is a positive constant, is the diffusion coefficient.
The feedback information Fi (t) created every fixed period τi by node i consists
of the following two quantities:
Node i reports this to the upstream node i − 1 with a period of τi = di−1 . Here,
the target transmission rate is determined as ri−1 (t) = Ji (1, t). Moreover, the
packet flow Ji (t) in node i is renewed whenever feedback information arrives
from the downstream node i + 1 (with a period of τi+1 = di ).
To enable an intuitive understanding, we briefly explain the physical meaning
of DFC. We replace i with x and apply a continuous approximation. Then the
propagation delay becomes di → 0 for all i and the flow (2) is expressed as
˜ x, t) = α r(x, t) − D ∂n(x, t) ,
J(α, (4)
∂x
and the temporal evolution of the packet density n(x, t) may be represented by
a diffusion-type equation,
3 Parameter Design
3.1 Approach
In DFC, there are two important parameters: one is the flow intensity multiplier
α and the other is the diffusion coefficient D. Our previous study shows that α =
1 is a natural and appropriate choice because that means the balance between
input and output traffic at a node. The residual problem is to determine an
appropriate value of D. The diffusion coefficient governs the speed of diffusion. In
physical diffusion phenomenon, larger D causes faster diffusion. If DFC model is
completely corresponding to physical diffusion phenomenon, a large value of D is
suitable for fast recovery from congestion. Unfortunately, DFC is not completely
corresponding to physical diffusion. As we see later in the next section, too large
value of D in DFC blocks diffusion phenomenon in networks. The reason of this
problem comes from the fact that networks have discrete configurations although
physical diffusion phenomenon occur in a continuous space-time environment.
That is, the spatial configuration of routers is discrete, and timing of control
actions is also discrete.
Conversely, too small value of D causes very slow diffusion, and this means
that stolid congestion recovery wastes much time.
Our approach to design a value of D is simple. We take a larger value of D
in the range of values in which diffusion can occur in networks.
The partial differential equation (5) describes temporal evolution of packet den-
sity in continuous approximation of networks. The first term on the right-hand
side in (5) describes a stationary packet flow and this is not concerned with
diffusion, but the second term is essential in diffusion. Thus, we consider the
following partial differential equation,
∂n(x, t) ∂ 2 n(x, t)
=D , (6)
∂t ∂x2
where this is the ordinary diffusion equation.
466 C. Takano, K. Sugiyama, and M. Aida
Of course, the structure of networks and the timing of control actions are not
continuous. Behaviour of DFC is described by a difference equation rather than
the differential equation. In other words, DFC make networks solve a difference
equation with discrete space x and discrete time t.
For simplicity, we assume all the links in networks have same length Δx. In
this situation, interval of DFC’s action is the same for all node, and we denote
it as Δt. The difference equation corresponding to (6) is as follows:
n(x, t + Δt) − n(x, t) n(x + Δx, t) − 2 n(x, t) + n(x − Δx, t)
=D . (7)
Δt (Δx)2
If the solution of (7) exhibit similar behavior to that of (6), DFC appropriately
works and diffusion of packet density occurs. Our issue is to find appropriate
value of D in which the solution of (7) exhibits diffusion phenomenon.
Let node position in 1-dimensional configuration be xk (xk+1 − xk = Δx; k =
0, 1, . . . , S), and time of DFC’s action be t (t +1 −t = Δt; = 0, 1, . . . , T ). We
take the boundary condition, n(t, x0 ) = n(t, xS ) = 0. If behavior of n(xk , t )
exhibits a diffusion effect with time,
lim n(xk , t ) = 0, (8)
→∞
We set the length of links as Δx = 1 and the interval of DFC’s control action (it
is equal to the propagation delay of a link) as Δt = 1, the range of the diffusion
coefficient D is
1
0<D< . (14)
2
Consequently, to make fast diffusion, we take a value of D as large as possible
in this range.
4 Simulation Results
In this section, we show simulation studies about the performance of DFC with
different values of the diffusion coefficient D in order to verify the range (14)
and our design policy of D. Simulations were made by using ns2 simulator [9].
We extended the simulation tool ns2 capability with the function of DFC.
100
0
1 15 29 1 15 29 1 15 29 1 15 29 1 15 29
node ID node ID node ID node ID node ID
300
number of packets that
100
0
1 15 29 1 15 29 1 15 29 1 15 29 1 15 29
are stored in each node
300
number of packets that
100
0
1 15 29 1 15 29 1 15 29 1 15 29 1 15 29
node ID node ID node ID node ID node ID
t = 0.1 s, respectively. After the background flow traffic entered the network,
the link from node 15 to 16 became a bottleneck, and traffic of both flows was
regulated by predefined rules for DFC. After congestion occurred, we investigated
the temporal evolution of the network state.
3000
total number of packets
2000
1000
0
0 0.5 1 1.5 2 2.5 time 0 0.5 1 1.5 2 2.5 time 0 0.5 1 1.5 2 2.5 time 0 0.5 1 1.5 2 2.5 time
D = 0.01 D = 0.40 D = 1.00 D = 3.00
Fig. 6. Temporal evolution of the total number of packets that are stored in nodes
that are in transit on links
1000
0
0 0.5 1 1.5 2 2.5 time 0.5 1 1.5 2 2.5 time 0 0.5 1 1.5 2 2.5 time 0 0.5 1 1.5 2 2.5 time
D = 0.01 D = 0.40 D = 1.00 D = 3.00
Fig. 7. Temporal evolution of the total number of packets that are in transit on links
Figure 6 denotes the temporal evolution of the total number of packets that
are stored in nodes, in cases of D = 0.01, 0.4, 1.0 and 3.0, respectively. The
horizontal axes denote simulation time and the vertical axes denote the total
number of packets that are stored in nodes. The first two cases are in the range of
(14). Both cases exhibit diffusion effect and the total number of packets decreases
with time, but it is very slow in the case of D = 0.01. This means that too small
D prevents fast recovery from congestion.
The last two cases in Fig. 6 are out of the range of (14). Larger value of
D > 1/2 causes instability of the total number of packets. In particular, too
large D exhibits chaotic behavior.
Next, we investigate how much packets are transmitted in the network. The
volume of packet transmission at time t can be denoted by the total number of
packets in transit on links of the network. Figure 7 shows the results in cases of
D = 0.01, 0.4, 1.0 and 3.0. The horizontal axes denote simulation time and the
vertical axes denote the total number of packets that are in transit on links.
Since the maximum number of packets in transit on a link at a moment was
100, and the background flow passed through about half of links of the target
flow, the maximum total number of target flow’s packets in transit on links
was 2, 900 when t ≤ 0.1 and was 1, 450 when t > 0.1. On the other hand, the
maximum total number of background flow’s packets in transit on links was 750
after t = 0.1. From the first two panels in Fig. 7, the numbers of packets in
transit on links for both flows reached almost their maximums in a short time
and these results mean that they fairly share the link bandwidth. For larger D
that is out of the range of (14), the total number of packets in transit on links
becomes unstable. These results show that larger value of D > 1/2 degrades the
performance of packet transmission.
These simulation results substantiate our design policy of D; to make fast
diffusion, we take a value of D as large as possible in the range of (14).
470 C. Takano, K. Sugiyama, and M. Aida
5 Conclusions
To overcome the difficulty in control of high-speed networks, we have proposed
DFC. In this control mechanism, the state of the whole network is controlled indi-
rectly through the autonomous action of each node; each node manages its local
traffic flow on the basis of only the local information directly available to it, by us-
ing predetermined rules. By applying DFC, the distribution of the total number
of packets in each node in the network becomes uniform over time, and it exhibits
orderly behavior. This property is suitable for fast recovery from congestion.
One of important issues in design of DFC is how to choose the value of diffusion
parameter. This is the central issue for enabling DFC to make diffusion effects
of packet density in networks. This paper investigates the appropriate value of
the diffusion parameter.
We determined the range of the diffusion parameter by applying the condition
for discrete space-time computations of the diffusion equation to DFC. On the
other hand, even if the value is in the range, too small value of the diffusion para-
meter causes very slow diffusion, and this means that stolid congestion recovery
makes to waste much time. Consequently, to make fast diffusion, we should take
a value of the diffusion parameter as large as possible in this range. Simulation
results verified our proposed design policy.
This research was partially supported by the Grant-in-Aid for Scientific Re-
search (S) No. 18100001 (2006–2010) from the Japan Society for the Promotion
of Science.
References
1. Y. Bartal, J. Byers, and D. Raz, “Global optimization using local information with
applications to flow control,” Proc. the 38th Ann. IEEE Symp. on Foundations of
Computer Science, pp.303–312, Oct. 1997.
2. S. H. Low and D. E. Lapsley, “Optimization flow control-I: basic algorithm and
convergence,” IEEE/ACM Trans. Netw., vol.7, no.6, pp.861–874, 1999.
3. J. Mo and J. Walrand, “Fair end-to-end window based congestion control,”
IEEE/ACM Trans. Netw., vol.8, no.5, pp.556–567, Oct. 1999.
4. R. Johari and D. Tan, “End-to-end congestion control for the Internet: Delays and
stability,” IEEE/ACM Trans. Netw., vol.9, no.6, pp.818–832, Dec. 2001.
5. C. Takano and M. Aida, “Stability and adaptability of autonomous decentralized
flow control in high-speed networks,” IEICE Trans. Commun., vol.E86-B, no.10,
pp.2882–2890, 2003.
6. C. Takano, M. Aida, and S. Kuribayashi, “Autonomous decentralized flow control in
high-speed networks with inhomogeneous configurations,” IEICE Trans. Commun.,
vol.E87-B, no.6, pp.1551–1560, 2004.
7. C. Takano and M. Aida, “Diffusion-type autonomous decentralized flow control for
end-to-end flow in high-speed networks,” IEICE Trans. Commun., vol.E88-B, no.4,
pp.1559-1567, 2005.
8. M. Aida, C. Takano and A. Miura, “Diffusion-type flow control scheme for multiple
flows,” The 19th International Teletraffic Congress, pp. 133-142, 2005.
9. The Network Simulator—ns-2.
http://www.isi.edu/nsnam/ns/
Bandwidth Management for Smooth Playback of Video
Streaming Services
1 Introduction
Recently, it has become usual that one can view a stored video file such as news clips
from high-speed Internet [5]. The bandwidth of a news clip in [5] with screen size of
6.5 by 5 inches is 492Kbps, from which we could witness frequent freezing of scenes
even for this narrow bandwidth video file due to limited bandwidth allocated to a
flow.
As the bandwidth of the down link for the end users in the current access network
increases from a few megabits per second (Mbps) to tens or hundreds of Mbps, we
envision that a need for a high-quality VoD (Video on demand) or IPTV (Internet
protocol TV) service increases in the near future.
If the bandwidth in an end-to-end path of a video stream is sufficient, the video file
will be played back smoothly after the initial buffering. However, when the network is
congested or if the bandwidth of a video server or user is not sufficiently provided, it
is inevitable that a degradation of the quality of service (QoS) will occur by repeated
freezing of the scenes due to frequent buffering. To resolve this problem, we have to
devise a method to manage bandwidth for a video stream so that a smooth playback of
a video file is provided to the users.
It is known that the delay of a sequence of packets transferred through the IP
network has a great spectrum of variations. The unpredictable behavior of packets
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 471 – 480, 2006.
© Springer-Verlag Berlin Heidelberg 2006
472 H. Lee, Y.K. Kim, and K.-H. Lee
inside the network causes a serious quality degradation problem to a receiving side of
a streaming service such as VoD or music. The random packet delay results in random
arrivals of packets at the receiving side, which causes buffer underflow or overflow.
Frequent packet underflow causes frequent buffering at the video decoder, and
consequently playback is not carried out smoothly. One can also note that some
packets are lost inside the network as well as at the receiving end when the buffer
overflows. Packet loss requires a retransmission of the lost packets from the source or
concealment of loss at the receiver, which disrupts smooth decoding of a video
stream.
Summarizing those two causes of QoS degradation for video streaming services in
IP network, we argue that a method to avoid the major sources of QoS degradation in
a packet-based video streaming has to be devised.
One can avoid this problem if one adopts a sophisticated service scheme such as
resource reservation at the access side as well as the differentiated packet scheduling
scheme architecture at the core of a network for a service with strict QoS
requirements. However, those service architectures require modification in the
network equipments at the network access and core, and so it is envisioned that those
service architectures will not be widely deployed in the near future.
One can think of CAC (connection admission control), adaptive video source
coding, dynamic bandwidth allocation or dimensioning method to guarantee QoS to
the users. There exist a lot of research works on the former two schemes [13]. Under
the current simple service structure of best effort IP network, it would not be efficient
to control the video source rate frequently based on the microscopic dynamism of
bandwidth consumption for each flow. Therefore, network operators may want a
simple network operation that relies not on the complex control of the associated
packets but on a network provisioning with an appropriate amount of bandwidth.
To that purpose, we will devise a bandwidth dimensioning and management
method to a video streaming service, because the loss and delay problem in packet-
based network that operates in best effort scheme can be resolved by provisioning at
least an appropriate bandwidth for a flow [11].
We argue that the delay and its variation of a packet in the network can be
minimized if we devise a traffic control scheme to the packet stream with random
inter-arrival time such that the packet arrival pattern of a video stream behaves almost
in packet train manner1.
In order to illustrate our argument in the context of IP network, we have two
purposes in this work. First, we propose a traffic control model called the packet
pacing scheme for IP access network and compare the delay performance between the
original random packet stream and the paced packet stream via approximate queuing
model. Second, we apply the model into the bandwidth dimensioning for the streamed
video delivery service network.
The contribution of the work is as follows: First, we propose a packet pacing
scheme in the video streaming server so that the video stream becomes smooth and the
mean delay becomes minimized. Via numerical experiment, we show that the delay
performance of paced packet stream is better than that of original random packet
stream by numerical experiment. Second, we propose a bandwidth management
1
Packet train means that consecutive flows of packets are almost equally distanced.
Bandwidth Management for Smooth Playback of Video Streaming Services 473
scheme for a video stream service by managing the buffer level so that neither buffer
starvation nor buffer overflow occurs. Via numerical experiment, we illustrate the
efficacy of the proposed method in the design of the optimal link capacity of video
server.
This paper is organized as follows: In Section 2 we describe the concept of pacing
the packet stream. In Section 3 we present a queuing model that represents the
expected queuing delays for the two packet streams, the randomly spaced packet
stream and the paced packet stream. In Section 4 we apply the proposed model to
managing the bandwidth for video streaming server. In Section 5 we present the result
of a numerical experiment and discuss the implication of our argument. Finally, in
Section 6, we summarize the paper.
IP Backbone Access
Video
server network
Finally, the access network of client in the receiving side may be composed of
DSL, E-PON (Ethernet passive optical network), or HFC (hybrid fiber and coaxial).
Fortunately, the traffic stream that flows down to the receiver from the access network
rarely meets collision due to the inherent broadcasting mechanism for the down
traffic. Therefore, the concern for the QoS of streaming video lies in the sending side
of the IP backbone network, specifically the video server side. As such, an appropriate
level of sustainable throughput has to be provided at the access network of a video
server in order that users can watch a video over the IP network with a satisfactory
QoS.
3 Delay Analysis
Under the prevalent environment of the current Internet with a slim pipe at the access
and a fat pipe at the core of the network, it can be easily found that no advantages are
obtained by having non-identical shapers for a flow at each node in an end-to-end path
of a flow in order to make smooth the traffic flow. Thus, it is recommended that
shaping is carried out only once at the ingress point of the network [6]. If we follow
this argument, the end-to-end delay of a packet is composed of the shaping delay and
the sum of nodal delays at the traversing nodes as well as the propagation delay and
some hardware processing delays at the sender and receiver. A detailed discussion
about the component of end-to-end delay is out of the scope of this work, and refer to
[10] for more information about it. We focus only on the buffering delay of the packet
pacer in order to investigate the efficacy of pacing to the expected waiting time of a
packet for a video streaming service.
Let us assume that the original packet stream is generated by Poisson distribution
with mean arrival rate λ and variance of inter-arrival time σA2. The service time of the
packet is generally distributed with mean service time 1/μ and variance σS2. The mean
offered load to the system is given by ρ=λ/μ. Let us define the squared coefficient of
variation for the inter-arrival time CA2 and the service time CS2 respectively, are given
by C A2 = λ2σ A2 and C S2 = μ 2σ S2 . Then, the original packet stream in the buffer of the
packet pacer is modeled as an M/G/1 queuing system with FIFO (First in first out)
packet services. Note that the mean delay performance of M/G/1 queuing system is
well-known [4], [8]. Using the result from [4] for the mean waiting time WO of an
arriving packet from the original packet stream, we obtain (1).
1 ρ2 1 + C S2 (1)
WO = ( × )
λ 1− ρ 2
Now let us discuss about the effect of pacing the original packet stream into an
equi-distant packet stream with inter-packet arrival time defined by d. The original
packet stream with Poisson arrival process is transformed to a general independent
arrival process with constant packet inter-arrival time d, where d = λ , so that the
1
effective arrival rate of the video source is the same for both models. Then, the paced
arrival process is D/G/1 queue with squared coefficient of variation for the inter-
arrival time CA2 to be zero.
Bandwidth Management for Smooth Playback of Video Streaming Services 475
Note that D/G/1 queue is a special case of GI/G/1 queuing system with fixed packet
arrival time. For the GI/G/1 queuing system, an approximation formula for the mean
waiting time of an arriving packet is given in [4]. If we modify the result of [4] by
changing the general independent arrival process to a constant inter-arrival time
process, we obtain the following result for the mean waiting time WP of the paced
packet.
ρ / μ C S2 2 1− ρ 1 (2)
WP ≈ × × exp( − . )
1− ρ 2 3 ρ C S2
If we compare the two formulas for WO and WP, the mean waiting time of M/G/1
queue and D/G/1 queue, respectively, we can find that the following inequality holds
if and only if ρ<1.
WP < WO (3)
Therefore, we can argue that the paced packet stream experiences smaller delay
than the original random packet stream. This is the first desired aspect of our
proposition.
2
Sustained bandwidth means the amount of bandwidth that can be provided continuously to a
flow by a network during the session duration time.
476 H. Lee, Y.K. Kim, and K.-H. Lee
Note that we can tune the offered load and the arrival rate of the traffic if the values
U and L are determined. On the other hand, the arrival rate of the traffic can be
controlled by tuning inter-packet gap d that we have defined in the previous section.
We can find that the formula (4) is a non-linear inequality, and we can obtain the
solution by using a numerical method. However, one can not see an explicit relation
between the system parameters and the acquired performance if one uses the
numerical method. Therefore, let us resort to an analytic approximation method so that
we can obtain a closed formula for the optimal sustainable load of the system, via
which we can obtain an intuition for the optimal load level of a system from assuming
a small number of parameters that we have defined up to now. From the preliminary
numerical experiment we found that the solution obtained by analytic approximation
method is closely matched with that obtained by numerical method.
Bandwidth Management for Smooth Playback of Video Streaming Services 477
1− ρ C S2
Let us define a variable X as X = , and let us also define χ as χ = . Then
ρ 2
we can rewrite (4) into (5) by using Maclaurin series and some manipulation, which is
given as follows.
1
χ− X
L≤ 3 ≤U (5)
X (1 + X )
Note that the formula (5) is composed of two inequalities: the left-hand inequality
(LHI) and the right-hand inequality (RHI), which are given as follows.
1
LHI :LX 2 + ( L + ) X − χ ≤ 0, (6)
3
1
RHI : UX + (U + ) X − χ ≥ 0
2
3
We can obtain a closed form solution for the inequalities (6), which is given as
follows.
( L + 1/ 3)2 + 4 L χ − ( L + 1/ 3)
XL ≤ , (7)
2L
(U + 1/ 3) 2 + 4U χ − (U + 1/ 3)
XR ≥
2U
Note that we can also obtain a range of an optimal load for the video delivery
service system from (7). Eq. (8) represents two loads, the upper bound and lower
bound of the load, which is represented by ρUP and ρ LP , respectively. In the sequel,
ρ YX stands for the load of a video stream, and the lower index X=L stands for the
lower bound and X=U stands for upper bound, and the upper index Y=O stands for the
original stream and Y=P stands for the paced stream.
1
ρUP = ,
1+ X R (8)
1
ρ LP =
1+ X L
Note from (7) and (8) that the upper and low bounds for the load of video traffic are
functions of the buffer threshold as well as the squared coefficient of variation for the
service time of the packet. The former parameter is controlled by a network operator,
whereas the latter one is generic to the traffic source. Note that we can estimate the
range of the required bandwidth for a video flow from (8).
As we have argued before, the two most effective ways to sustain a satisfactory
level of the QoS for the video services is as follows: First, shape the arrival pattern of
the video packets to be constant by packet pacing. Second, keep the range of the
offered load of the video stream between ρ LP and ρUP . In this respect, the above result
gives us a very useful insight for the provisioning of a video streaming service.
478 H. Lee, Y.K. Kim, and K.-H. Lee
5 Numerical Results
First, let us compare the range of the offered load for the original video stream and
paced video stream, and verify the validity of our argument in (3). Note that we have
presented the load for the paced video stream in (8). The upper and lower loads, ρUO
and ρ LO , respectively, for the original video stream can be also obtained in the same
analogy that we have used in deriving the result (8), which is shown in (9).
L2 + 4η L − L
ρ LO = ,
2η (9)
U + 4ηU − U
2
ρUO =
2η
1 + C S2
where η is defined by η = .
2
In order to compare the performance between the original video stream and the
paced video stream, we carry out the experiment by assuming the same threshold
parameters for the buffer as well as the traffic sources. Let us assume that the average
number of frames generated by a video source is 15frames per second. A video frame
is represented by 250×200pixel, and a pixel is represented by one byte. Then, the
mean data rate of a video source is 6Mbps, which is denoted by r = 6Mbps.
As to the packet traffic model, let us assume that a video frame is composed of a
number of packets with Gaussian distributed packet length. We use the traffic source
parameters that have been discussed in [7]. The mean packet length is assumed to be
50bytes and maximum packet size of 125bytes, and the standard deviation of the
packet length is 25bytes. The lower limit L of the buffer is assumed to be one packet,
so that L=1, while the upper limit U of the buffer is assumed to be a design parameter,
which varies from 2 to 47.
Fig.2 illustrates the sustainable load of the system, which is an upper bound of load.
One can find that the upper limit on the offered load of a paced video stream is higher
than that of original video stream, which means that the delay performance of paced
stream is better than that of original stream, which complies with our first argument.
1.2 12
Bandwidth [Mbps]
1 10
Offered load
0.8 8
0.6 6
0.4 Paced stream 4 Paced stream
0.2 Original stream 2 Original stream
0 0
2 7 12 17 22 27 32 37 42 47 2 7 12 17 22 27 32 37 42 47
Upper threshold Upper threshold
From the numerical experiment, we found that the lower bound of the offered load
is ρ L =0.5 for the paced stream and ρ L =0.46 for the original stream, which correspond
P O
to the minimum load that has to be sustained in the system so that buffer starvation
does not occur. This means that the offered load of a playback buffer has to be kept no
smaller than 0.5 for the paced stream and 0.46 for the original stream, otherwise the
buffer suffers starvation. From the above result, we can also estimate the required
bandwidth CP for a paced video stream as well as CO the original video stream under
the assumed system parameters. The result is given in Fig.3.
We found that, in order to avoid the buffer starvation and overflow, the bandwidth
of a video server has to be maintained between C LP =6.25Mbps and CUP =12Mbps,
respectively, for a video flow with packet pacing, whereas the bandwidth of a video
flow without packet pacing has to be maintained between C LO =6.30Mbps and
CUO =13.04Mbps. Here, the lower index L and U stands for lower and upper bound,
respectively. From this result we obtain the following conclusions: The smoother the
traffic stream, the smaller the required bandwidth. This is the final finding of this
work.
From Fig.3 one can also find that the required bandwidth of a video stream
decreases as the upper threshold of a buffer increases. Therefore, we can conclude that
there exists a trade-off between the required bandwidth and the buffer capacity for a
video stream service, and one has to prepare a sufficient buffer space, say about a few
tens of packets, if one wants to avoid over-provisioning of bandwidth to a video
streaming service.
Note that one has to focus more on the lower limit of the bandwidth, because the
lower limit of the bandwidth is the minimum bandwidth that has to be sustained to a
video stream so that the video frames are played back in a smooth manner without
starvation in the playback buffer. On the other hand, the upper limit of the bandwidth
can be used as a reference for determining the size of the buffer.
6 Conclusions
In this work we proposed a packet transfer control and bandwidth management
scheme for a smooth playback of a stored video stream over IP network. To that
purpose, we first argued that the packet pacing guarantees a minimum delay to
packets, whereas the bandwidth management scheme guarantees smooth playback of
video frames. By using queuing model and numerical experiment, we validated our
argument. Next, we proposed a buffer management scheme for guaranteeing smooth
playback of streamed video data, and we obtained an explicit formula for the optimal
level of offered load for a video streaming service. Finally, we presented a
quantitative result for the manageable range of bandwidth for sustained quality of
video streaming service by assuming typical parameters for the traffic source.
The intuition obtained from this work is that a better quality for the video stream
can be provided by sustaining an optimal level of bandwidth to a video flow. We also
showed that the delay of packets from the video source can be reduced a priori by
pacing the packets from the video server when the backbone network is assumed to be
transparent of delay to the transit packets. The result can be also applied to the
480 H. Lee, Y.K. Kim, and K.-H. Lee
Acknowledgments
The authors are grateful to the anonymous reviewers for their helpful comments,
which improved the presentation of the paper.
References
1. Acharya S., Smith B., Parns P., Characterizing user access to video on the World Wide
Web, Proc. ACM/SPIE MMCN Jan. 2000.
2. Apostolopoulos J.G., Video communications and video streaming, Streaming media
systems group, Hewlett-Packard Laboratories, May 1, 2000.
3. Besset C., Le Drogo C., Dumetz C., Paquette R., Orange video project with Alcatel,
Alcatel Telecommunications Review 4th Quarter, 2005.
4. Bolch G., Greiner S., de Meer H., Trivedi K., Queueing Networks and Markov Chains,
John Wiley & Sons, Inc. 1998.
5. http://www.chosun.com/tv/news/
6. Georgiadis L., Guerin R., Peris V., Sivarajan K., Efficient network QoS provisioning based
on per node traffic shaping, IEEE/ACM Transactions on Netwoking, Vol.4, No.4, August
1996.
7. Kim D.-H., Jun K., Dynamic bandwidth allocation scheme for video streaming in wireless
cellular networks, IEICE Trans. Commun., Vol. E89-B, No.2, February 2006.
8. Kleinrock L., Queuing systems, Volume 1: Theory, John Wiley & Sons, 1975.
9. Komori Y., Kasahara S., Sugimoto K., A study on dynamic rate control mechanism based
on end-user level QoS for streaming services, Technical Report of IEICE NS 2003-332
(2004-03).
10. Lee H., Back Y.-C., Anatomy of delay for voice services in NGN, Proceedings of Fall
Conference of the Communication Society of IEEK, 2003, Korea.
11. Lee H., Sohraby K., Flow-aware link dimensioning for guaranteed-QoS services in
broadband convergence networks, Paper submitted to JCN, March 2006.
12. Sivaraman V., Chiussi F., Gerla M., Traffic shaping for end-to-end delay guarantees with
EDF scheduling, Proceedings of IWQoS 2000.
13. Wu D.P., Hou Y.W., Zhang Y.Q., Scalable video coding and transport over broadband
wireless networks, Proc. IEEE, 2001, 89.
14. Le Boudec J.-Y., Network calculus made easy, Technical report EPFL-DI 96/218, Dec.
1996.
15. Sardella A., Video transit on an MPLS backbone, A solution brief from Juniper networks,
Juniper Networks, Inc., 200106-0
An Enhanced RED-Based Scheme for
Differentiated Loss Guarantees
1 Introduction
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 481–490, 2006.
c Springer-Verlag Berlin Heidelberg 2006
482 J. Koo, V.V. Shakhov, and H. Choo
2 Related Work
We briefly review the previous works in this section. As mentioned earlier, many
related works have been proposed to meet the demand for differentiated loss
guarantees in the future Internet. Specifically, the proportional loss differentia-
tion (PLD) model is one of such efforts, which offers fixed proportions on loss
rates between the QoS levels of the different classes rather than absolute bounds.
Dovrolis et al. [4] who developed first the PLD model, claim that several pre-
vious mechanisms such as complete buffer partitioning (CBP), partial buffer
sharing (PBS), or multi-class RED, are not suitable for relative differentiated
services. They propose and evaluate Proportional Loss Rate (PLR) mechanisms
An Enhanced RED-Based Scheme for Differentiated Loss Guarantees 483
that closely approximate the PLD model. Liebeherr et al. [7] propose a novel
algorithm called Joint Buffer Management and Scheduling (JoBS) for the inte-
gration of buffer management and packet scheduling and compare with PLRs
for PLD service. Li et al. [8] propose a novel algorithm by employing Probabilis-
tic Longest Queue First (PLQ) mechanism and claim that its implementation
cost is much less than PLRs and even more practical. Zeng et al. [10] propose
more enhanced dropping algorithm than PLRs in terms of the packet shortage
phenomenon.
More recently, the Bounded Random Drop (BRD) proposed in [6] offers ser-
vice differentiation while keeping the complexity low. The performance goals of
its scheme are to achieve both absolute and relative loss requirements without
introducing too much added complexity such as implementation, configuration,
and deployment. Since BRD is implemented using a single FIFO queue and a
random dropping mechanism, it is simpler than JoBS in complexity perspective.
In addition, BRD demonstrated the possibility of providing per-hop differen-
tiated loss guarantees without additional active management such as policing,
traffic profiles, or signaling. Moreover, while JoBS shows significant deviations
in the desired short timescale, BRD is capable of providing both long and short
term performance guarantees.
Compute
Estimate Packet Path
Drop
Input Rates
Probabilites
Control Information Path
Packet
Arrival Random
Classifier FIFO
Dropper
(1) Arrival rates are estimated using an exponentially weighted moving average
with a weighting factor α. For each class i, we use a counter Ai to keep track
of the amount of input traffic during each Δt sampling period. At the end
of each period, the traffic rates are updated by ri = (1 − α)ri−1 + αAi /Δt,
for class i = 1, ..., N .
(2) The target loss probabilities, pi , i = 1, ..., N , are computed based on the ri ’s
and all counters are reset.
(3) Upon arrival of a packet belonging to class i, the packet will be dropped
randomly with the calculated loss probability of class i.
Although BRD has several benefits such as low complexities and good func-
tionalities, we identify that it has some shortcomings (as shown in section 3).
Next, we focus on the BRD scheme as it is more relevant to our work.
484 J. Koo, V.V. Shakhov, and H. Choo
3 Proposed Scheme
3.1 Motivation
We first analyze the target loss probability in BRD using the equations presented
in [6] and show the result in Table 1. For this analysis, we consider a single output
link with capacity 10 Mbps. We assume three classes with Constant Bit Rate
(CBR) sources. The loss bounds assigned to each class are set to 0.1 for Class
1, 0.2 for Class 2, and none for Class 3. The input rates of the three classes are
shown in Table 1.
As we can see, all packets are not discarded for low traffic intensity (see Cases
1 and 2 in Table 1). Case 3 shows that the higher priority classes (Classes 1
and 2) experience the same loss performance as lower priority ones as long as
their absolute loss bounds are not violated. Case 4 shows that the higher priority
classes (Classes 1 and 2) receive preferential loss treatment only when they are
required to avoid violating their own loss bounds. From Cases 5 to 8, we know
that the loss bounds of lower priority classes are relaxed first when it is not
feasible to satisfy the loss bounds of all traffic classes simultaneously.
Through the analysis of the target loss probability in BRD, we identify that
BRD has several shortcomings as the following:
• Long Queuing Delays - The BRD scheme does not drop packets before
traffic load is greater than one (see Cases 1 and 2 in Table 1). Therefore,
it exists occasionally in a case that a buffer is full or reaches a specified
threshold, thereby resulting in long packet delays in the buffer.
• Fairness - The BRD scheme allows a higher priority class or a few higher
priority classes to monopolize the buffer space of the router (see Cases 6-8
in Table 1), preventing other lower priority classes from getting space in the
router queue. If the higher priority classes continue to arrive with large traffic
volume for a long period of time, the lower priority classes will be highly
likely starved. This results in unfair sharing of network resources among the
An Enhanced RED-Based Scheme for Differentiated Loss Guarantees 485
classes, thereby giving rise to fairness problems. Moreover, 100% drop of the
lower priority classes causes all their senders to back off simultaneously -
called global synchronization problem.
• Computational Complexity - The BRD scheme needs just one compar-
ison for each incoming packet (refer to the BRD algorithm given in the
previous section). Moreover, after each sampling period, a few operations
including the random number generation are required. They increase as the
number of incoming packets increases.
• Selection of Optimal Values of Parameters - The BRD scheme depends
on the values of its two parameters, sampling period Δt and a weight factor
α. Therefore, it is necessary to investigate further the sensitivity of optimal
values of Δt and α.
The shortcomings above are investigated for the deterministic traffic. We also
examine that the BRD scheme does not support declared level of QoS for the
stochastic traffic. Actually, let a size of considered buffer be equal to K. Let pi be
the probability of i packets in the buffer, i = 0 . . . K. Here pK is the probability
of buffer overflow. At this time, all incoming packets are dropped. According to
the BRD algorithm, if the number of packets in the buffer is equal to i, where
i = 1 . . . K − 1, then the drop rate of offered load equals LB1 , where LB1 is
the required level of blocking probability in Class 1. In the case of K packets in
the buffer, the blocking probability of incoming packets equals one because the
considered buffer is full.
Let us calculate the drop rate, DR, of the higher priority packets
K−1
DR = LB1 pi + pK · 1 (1)
i=0
K
= LB1 pi + (1 − LB1 )pK (2)
i=0
= LB1 + (1 − LB1 )pK > LB1 . (3)
K
where i=0 pi = 1 and LB1 < 1. Thus, the drop rate of offered load corresponds
to the required level of QoS if and only if pK = 0. It means unlimited buffer,
cut-through switching under deterministic traffic, or very low traffic intensity.
Otherwise, DR > LB1 and this indirectly implies that the BRD scheme does not
support the required level of QoS. It may be concluded that ignored buffer status
information is certain to be a shortcoming in QoS-enabled schemes. Accordingly,
one of the research objectives of this paper is to eliminate the weaknesses dis-
cussed here and to present strong service guarantees with low complexity and
high scalability.
3.2 MQDDP
(1) packet classifier that can distinguish packets and group them according
to their different requirements;
(2) packet dropper that determines both of the following: how much queue
space should be given for certain kinds of network traffic and which packets
should be discarded during congestion;
(3) packet scheduler that decides the packet service order so as to meet the
bandwidth and delay requirements of different types of traffic; and
(4) calculation module for the maximum drop probability that takes the
real-time measurement of traffic load and manages the packet dropper.
We first present the basic algorithm of RED on a single queue, and then we
extend to the proposed scheme by deriving the drop probability equations based
on a queueing model. Since each class-based queue in the proposed scheme is
independent and has common properties of RED, the derived equations below
are immediately applicable to each queue.
An Enhanced RED-Based Scheme for Differentiated Loss Guarantees 487
For a queue with RED, incoming packets are dropped with a probability that
is an increasing function d(k) of the average queue size k. The average queue
size is estimated using an exponentially weighted moving average formula. Let
k0 be the previous average queue size, then k = (1 − wq ) · k0 + wq · k, where
k is the current queue size and wq is a weight factor, 0 ≤ wq ≤ 1. RED offers
three control parameters: maximum drop probability maxp , minimum threshold
minth , and maximum threshold maxth . It depends on the averaged queue length
k with weighted factor wq to tune RED’s dynamics.
For a given resource portion out of common output channel through a predes-
tined scheduling, the service rate of each class queue i, μi , is determined. With
a certain arrival rate, λi , the related system utilization factor, ρi = λi /μi , can
be assumed. The typical dropping function of class i, di (k), in RED is defined
by three parameters minth , maxth and maxp,i as follows:
⎧
⎨ 0, k < minth
maxp,i ·(k−minth )
di (k) = maxth − minth , minth ≤ k < maxth (4)
⎩
1, k ≥ maxth
where maxp,i is maximum drop probability of Class i. Therefore, the drop prob-
ability of a packet depending on the dropping function related to each state k
can be calculated as follows:
K
Pi = πi (k)di (k), minth ≤ k < K (5)
k=minth
(1) Traffic loads and current queue sizes are monitored on the tth sampling
interval.
488 J. Koo, V.V. Shakhov, and H. Choo
0.07
0.06
0.05
0.04
Loss Probability = 0.02
0.03
0.01
1.2 1.25 1.3 1.35 1.4 1.45 1.5
Offered Load (ρ)
(2) At the end of each period, the values of maxp,i are obtained from the maxp -
by-ρ reference table for loss requirements and average queue sizes are esti-
mated by k.
(3) di (k) and πi (k) are computed by Eq. (4) and Eq. (6), respectively.
(4) The loss probabilities, Pi , i = 1, ..., N , are calculated by Eq. (5).
(5) Upon arrival of a packet belonging to class i, the packet will be dropped
randomly with the calculated loss probability of class i.
It is certain that the packet drop rate of the proposed method is a continuous
function of outgoing channel capacity. Hence, we can obtain more reasonably
the required level of packets blocking probability by rescheduling the outgoing
channel bandwidth. Thus, the proposed scheme provides absolute differentiated
loss guarantees. For example, if maxth = 64, minth = 2, maxp = 0.5, wq = 1,
Poisson flow of incoming packets, and the required level of packets loss is equal
to 10%, then the capacity of outgoing channel should be greater than 90% of
the offered load rate. If the required level of packets loss is 5%, then we need to
increase the outgoing channel bandwidth up to the offered load rate.
0.8
BRD
MQDDP
0.7
0.6
Packets Drop Rate
0.5
0.4
0.3
0.2
0.1
1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2.0
Offered Load Intensity/Outgoing Channel Capacity
Fig. 4. Comparison of BRD and MQDDP under very high offered loads
An offered load is assumed to be the Poisson process with rate kC, where
C is the outgoing channel capacity, k is a constant, and 1.2 ≤ k ≤ 2. Let the
maximal number of packets in the buffer be 64. The parameters of the proposed
scheme are given by maxth = 64, minth = 2, maxp = 0.5, and wq = 1.
According to the BRD algorithm, a part of the offered load is randomly
dropped. In the case of highly offered loads, the probability of incoming packet
dropping equals 1 − k1 . The packet drop rate for BRD scheme is
1 − pBO
1− (7)
k
where pBO is the probability of buffer overflow. It has been defined by the block-
ing probability formula of M/M/1/64 queuing system.
By assigning different offered load rates, we can get packet drop rates for each
scheme as shown in Figure 4. It may be concluded that the proposed scheme
has better performance in terms of packet drop rate. For instance, if the ratio
of offered load intensity to outgoing channel capacity equals 1.5, the proposed
scheme reduces the packets drop rate about 40%.
5 Conclusion
original RED with a mechanism for absolute loss guarantees, eliminates the men-
tioned weaknesses as shown the comparison between MQDDP and BRD in case
of high traffic intensity, and is shown to have the better performance in terms of
packet drop rate. In the future, it is reasonable to investigate an optimal choice
of several parameters in the proposed scheme with differentiated services.
References
1. J. Aweya, M. Ouellette, and D. Y. Montuno, “Weighted Proportional Loss Rate
Differentiation of TCP Traffic,” Int. J. Network Mgmt, pp. 257-272, 2004.
2. Y. Chen, M. Hamdi, D. Tsang, and C. Qiao, “Proportional QoS provision: a uni-
form and practical solution,” In Proc. of ICC 2002, pp. 2363-2366, 2002.
3. C. Dovrolis and P. Ramanthan, “A Case for Relative Differentiated Services and
the Proportional Differentiation Model,” IEEE Network, 13(5):26-34, 1999.
4. C. Dovrolis, D. Stiliadis, and P. Ramanathan, “Proportional Differentiated Ser-
vices: Delay Differentiation and Packet Scheduling,” In Proc. of ACM SIGCOMM,
pp. 109-120, 1999.
5. C. Dovrolis and P. Ramanathann, “Proportional differentiated services, part II:
loss rate differentiation and packet dropping,” In Proc. of IWQoS, pp. 52-61, June
2000.
6. Y. Huang and R. Guerin, “A simple FIFO-based scheme for differentiated loss
guarantees,” In Proc. of IWQoS, pp. 96-105, 2004.
7. J. Liebeherr and N. Christin, “JoBS: joint buffer management and scheduling for
differentiated services,” In Proc. of IWQoS, pp. 404-418, 2001.
8. J-S Li and H-C Lai, “Providing proportional differentiated services using PLQ,”
In Proc. of Globecom, pp. 2280-2284, 2001.
9. Vladimir V. Shakhov, Jahwan Koo and Hyunseung Choo, “On Modelling Reliabil-
ity in RED Gateways,” In Proc. of ICCS, pp. 948-951, 2006.
10. J. Zeng and N. Ansari, “An enhanced dropping scheme for proportional differenti-
ated services,” In Proc. of ICC, pp. 1897-1901, 2003.
Dynamic Location Management Scheme Using Agent in a
Ubiquitous IP-Based Network
1 Introduction
As all countries of the world are trying to construct ubiquitous infra, the present 3rd
generation network is ready to evolve into 4th generation network, in which IP is a
backbone network. Since the advent of ubiquitous age will bring about an explosive
increase of large multimedia traffics, ubiquitous networks should provide wideband
seamless mobility and stabilized services simultaneously. ITU-R has presented that
every telecommunication networks would be changed to IP based networks to
transmit large multimedia traffics with stability. [1]
NTT DoCoMo has suggested IP based IMT Network Platform (IP2) [2],[3] as the
next generation All-IP mobile network structure taking into account the increased
multimedia traffics and IP technologies. The basic structure of IP2 is categorized into
three classes, which are IP-Backbone(IP-BB), Network Control Platform(NCPF) and
Service Support Platform(SSPF) respectively. Among Mobility Management
functions of NCPF, Local Manager(LM) plays an important role to manage
information of all MN locations and enable the MN in Dormant state to change its
state to Active after conducting Paging process. During the Paging process, LM
makes flooding the Paging packets to every node in LA (Location Area) to which MN
is belong. At the moment, the overall network efficiency is significantly lowered to
*
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 491 – 500, 2006.
© Springer-Verlag Berlin Heidelberg 2006
492 S.-Y. Shin et al.
process the Paging of one MN resulting in the waste of network resources. To resolve
this problem, we suggest using an agent server for Paging. The agent server for
Paging will have the detailed information of MN location so that it mitigates the
waste of network resources by conducting selective Paging at the request of Paging.
In Chapter 2, the procedures of the conventional IP2 network Paging and the
proposed Paging mechanism with agent are described. The mobility of IP2 network is
described in Chapter 3 respectively. Paging Cost in IP2 network with agent server is
described in Chapter 4 and Network simulation results in Chapter 5 and conclusions
in Chapter 6 are presented.
Fig. 1. State transition diagram Fig. 2. Regional Classes of Location Area and
Routing Area
LM manages the location information of all MNs within IP2 networks and changes
MN in Dormant state to Active state by Paging procedure. Besides, it identifies the
area, at which MN is located, as Location Area Address(LAA) and stores it in
Location Management Table. RM assigns IPra to active MN and manages the
information of routing addresses. AR changes IPha in the packets into IPra and
transmits the packets to the network. AR has CST, which stores IPha and IPra of
Dynamic Location Management Scheme 493
source MN, and CDT, which stores IPha and IPra of destination MN [5]. LM
manages the information of MN locations in terms of LA. And RM manages the
information of MN locations concerning with RA. Figure 2 shows the regional classes
of LA and RA.
Since there is no concept of dormant state in Mobile IP, the location information
should be updated every time MN moves even if there is no communication.
Consequently, the waste of power and network resources becomes significant [6]. In
IP2, however, the problem of wasting power and network resources are mitigated
because LM maintains a dormant state of MN. The new problem of using LM is that
the excessive use of network resources is caused when the state is changed from
Dormant to Active through the Paging procedure, by which the messages are flooding
to all MN within LA. To resolve this new problem, we proposed to use an agent
server for Paging. Since the agent server has more detailed information of MN
location than LM does, it can obtain the information of a specific path while the
Paging messages transmitted from LM pass through the agent server. As a result, only
AR and Base Station(BS), to which IPha of MN belongs, would be transmitted the
Paging packets and the network resource usage becomes more efficient.
Figure 3 shows the procedure of registering detailed information of MN location
on an agent server. The location register is the procedure of registering MN’s new
location in case that MN joins a certain LA first or moves to other LA. In general, the
location registration is conducted during MN’s moving to other LA. The sequence of
the location register is as follows. 1) MN#1 in Dormant state detects its location has
changed. 2) MN#1 transmits a location registering message to BS3 of AR2. 3) AR2
transmits a location update message to LM. 4) Entry of MN#1 in the location
management table of LM is updated. 5) The location registering message sent by AR2
is transmitted to the agent server. 6) Entry of MN#1 in the agent table of the agent
server is updated. 7) The agent server transmits a location updating Ack to AR2. 8)
AR2 transmits Ack about the location registering message to MN#1. A location
registering procedure is finished in this way. Figure 4 shows the Paging procedure in
general IP2 networks without agent servers. It can be noticed that the resources are
Agent Table
Agent Server
A g e n t T a b le
Agent
S e rv e r
Figure 5 shows the Paging procedure in IP2 network environment with agent
server. Since the agent server has the exact information of AR and BS, at which
MN#1 is placed, it is not needed to transmit Paging messages to full area of LAA.
Dynamic Location Management Scheme 495
Agent
Server
Location Registration
Procedure
Paging
Procedure Buffering
IPra Allocation
Data Transmission
Fig. 6. The message flow in IP2 network (without agent server(left) / with agent server(right))
After AR receives the data packet for MT#M transmitting to MT#C, it finds some
IPra information about the destination of data packet at it’s own DST. After assuring
that AR does not have IPra for destination node, MT#C, it requests IPra for MT#C to
RM. The absence of IPra for MT#C causes Paging request to LM to find the exact
location of destination node, MT#C. After LM searches LA for IPha of destination
node, it floods Paging message to LA having MT#C. After MT#C receives this Paging
message, AR assigns new IPra to MT#C through Activation procedure and registers
new IPra to RM. At the next step, RM forwards IPra of MT#C to source AR which
MT#M is belonged. The data packets buffered at source AR are send to MT#C
following after substitution IPra for destination address of packets.
Figure 7 explains the message flow diagram for location register in IP2 network
environment with agent server. At the location register process, LM delivers message
for location information to agent server. In the course of Paging process, LM sends
Paging request message to agent server. In other words, the agent server acts for
Paging mechanism.
496 S.-Y. Shin et al.
Agent Server
Agent Server
In case of 1), sender MN#M in IP2 domain 3 transmits the first packet for
communication to receiver MN#C in AR1 of IP2 domain 1. Then AR6 makes a search
for IPra corresponds to the destination IPha of MN#C in its CDT. Since IPha of
MN#C is not listed in the table directory, however, AR6 asks RM in MM3 for IPra
with regard to IPha of destination MN#C. If AR6 finds there is no IPra in RM in MM3
also, AR6 checks on if exists in its LM in Dormant state. If AR6 finds it doesn’t exist,
AR6 asks MMs in other domain for IPha information. After AR6 finds the IPra does
exist in MM1 of domain 1 in dormant state, Paging procedure is conducted in LM of
MM1. Then, LM asks agent servers to conduct Paging to the location where MN#C is
located. Through the Paging procedure, MN#C conducts activation procedure and
AR1 allocates IPra and registers it to RM of MM2. This IPra information is
transmitted to AR6 in IP2 domain 3. Then, IPra in CDT of AR6 is updated and the
packet changes its destination address into IPra and is transmitted to the destination.
In case of 2), MN#C, which is communicating with MN#M, moves from LAA:a to
LAA:b region. During this stage, the location register procedure should be conducted
because the handover is a guarantee to mobility. Through the location register
procedure, LM, a location information of MN#C in agent servers and the information
of CDT in AR, at which every communicating MN is located, are updated. The
situation of 3) is that MN#C is moving within LAA:b region, allocated new IPra from
AR and registers it to RM. Then, RM updates CDT information of AR, at which all
Dynamic Location Management Scheme 497
MN communicating with MN#C is located. Lastly the case of 4) is that MN#C moves
from domain 1 to domain 2. MN#C conducts a location register procedure and
registers its IPha to LM of MM3. The location information of MN#C is deleted in LM
and RM of MM1. Then, IPra is allocated to AR4 and RM of MM2 CDT information of
AR4, at which all MNs communicating with MN#C are located, is updated.
(4)
(5)
5 Simulation
For the performance analysis of the proposed agent server, which aims to more
efficient Paging in IP2 environment, we have conducted a series of simulation using
NS-2 (Network Simulator 2). There is no conventional package perfectly supporting
IP2 environment in NS-2. CIMS(Columbia IP Micro-Mobility Suite), which is a NS-2
extension, supports micro mobility-related Cellular IP, Hawaii, Hierarchical Mobile
IP [10]. Among these three packages, Cellular IP has some features of IP2. In this
package, IP is used to discriminate hosts and MN has active / dormant state. Besides,
MN has features supporting the location management and Paging functions.
Fig. 10. Paging Load (BS:8, MN:2) Fig. 11. Paging Load (BS:8, MN:8)
Table 1. Paging Load (BS:8, MN:2) Table 2. Paging Load (BS:8, MN:8)
Simulation results shows that the performance of the proposed technique increases
as the area of LA becomes larger, as the number of BS increases and as the number of
MN increases. The maximum improvement rate was 45 %. Since the area of LA and
the MN node number is on an increasing trend in real world backbone networks, the
proposed method is expected to be more effective one, especially in the next
generation network environment, in which the number of ubiquitous terminals would
be increased rapidly.
6 Conclusion
As a 4G (Generation) network architecture, IP2 is the architecture which is based on a
new concept and proposed to cope with large multimedia traffics generated by mobile
communications in the near future. MN in IP2 has either dormant or active state.
Dormant state is a suspended state of communication and the state has to be changed
into active state through Paging procedures for restarting communications. During this
Paging procedure, signaling messages have to be transmitted to all terminals within LA.
Consequently, sharing resources of network is used excessively since many MNs shall
ask communication during dormant state or handover between LA. To mitigate this
shortcoming of the Paging feature of LM in IP2, a scheme of using agent servers is
500 S.-Y. Shin et al.
proposed in this paper. Since the agent servers for Paging procedures are keeping the
information of detailed location of MN, the Paging message can be transmitted to the
exact location of LA in case of a Paging request from LM. Consequently, the sharing
resources of the network can be used efficiently. NS-2 simulation results for the
performance evaluation of the proposed technique have shown that the Paging
efficiency can be improved by up to 45 %. Since the proposed technique does not
support any registration procedure according to handover in case that dormant MN
moves to other LA, however, additional supplement of this feature is needed.
Acknowledgement
This research was supported by the MIC(Ministry of Information and Communica-
tion), Korea, under the ITRC(Information Technology Research Center) support
program supervised by the IITA(Institute of Information Technology Assessment).
Reference
1. ITU-R Draft Recommendation, “Vision, framework and overall objectives of the future
development of IMT-2000 and systems beyond IMT 2000,” November 2002
2. H.Yumiba,et al., “ IP-based IMT Network Platform,” IEEE Personal Communication
Magazine, Vol. 8, No. 5, pp. 18-23, October 2001.
3. K. Imai, M. Yabusaki, and T. Ihara, “IP2 Architecture towards Mobile Net and Internet
Convergence,” WTC2002, September 2002.
4. Atsushi IWASAKI, Takatoshi OKAGAWA, “Scalability Evaluation of IP-based IMT
Network Platform.,” IEICE Technical Report, Vol. 104, No. 184, NS2004-69, pp. 5-8, July
2004.
5. Takatoshi Okagawa, et al., “Proposed Mobility Management for IP-based IMT Network
Platform,” IEICE Transactions on Communications 2005, Vol. E88-B , No. 7, pp.
2726~2734, 2005
6. Katsutoshi Nishida, et al., “Implementation and Evaluation of a Network-Controlled
Mobility Management Protocol (IP2MM): Performance Evaluation Compared with Mobile
Ipv6,” Wireless Communications and Networking Conference, 2005 IEEE, Vol. 3, pp.
1402 – 1408, March 2005
7. C. Rose and R. Yates, “Ensemble polling strategies for increased paging capacity in
mobile communication Networks,” ACM-Baltzer J. Wireless Networks, vol. 3, no. 2, pp.
159-177, 1997
8. Dong-Jun Lee, et al., “Intelligent Paging Strategy based on Location Probability of Mobile
Station and Paging Load Distribution in Mobile Communication Networks,” 2004 IEEE
International Conference on, Vol. 1, pp. 128 - 132. June 2004
9. E. Del Re, Senior Member, IEEE, R. Fantacci, G. Giambene, “Handover and Dynamic
Channel Allocation,” IEEE Transactions ov vehicular technology, VOL. 44, NO. 2, 1995
10. A. T. Campbell, Gomez, J., Kim, S., Turanyi, Z., Wan, C-Y. and A, Valko “Comparison of
IP Micro-Mobility Protocols,” IEEE Wireless Communications Magazine, Vol. 9, No. 1,
February 2002.
Detecting and Identifying Network Anomalies by
Component Analysis
1 Introduction
Nowadays, computer networks and traffic running through them are increasing at a
high pace to meet users' requirement. Beside the major proportion of productive
traffic, there are many types of network anomalies. The prominent point of all
network anomalies is that they generate abnormal changes in traffic features such as
bandwidth, load, and other traffic metrics. In this paper, we concentrate on large
network topologies connecting many networks by multiple links. Controlling
network anomalies in this scope requires collecting traffic and offline processing
data on all network links simultaneously. In our research, we propose to use simple
link utilization metrics, i.e. bandwidth (bps), load (packet per second), and counters
for selective packets to detect and diagnose network anomalies. This paper applies
component analysis and uses subspace method to detect abnormal exhibitions in
each link metric. This discovery about anomalous features in network traffic allows
us to detect and identify them in a timely manner. The efficiency of this method
depends on the data sampling rate which contains the detailed level of network
traffic.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 501 – 504, 2006.
© Springer-Verlag Berlin Heidelberg 2006
502 L.T. Quyen, M. Zhanikeev, and Y. Tanaka
IGMPFlood
UDPFlood
16 12
12 6
8
0 4
0 2 4 6 8 10 12 14 16 18 20 22 24
0
9 x10 4
Flash
10
6
8 3
6 0
Destination
90
-socket
4
60
2 30
0 0
0 2 4 6 8 10 12 14 16 18 20 22 24 0 2 4 6 8 10 12 14 16 18 20 22 24
Time (hour) Time (hour)
(a) Residual bandwidth vector (b) Residual vectors of network
at different sampling intervals link metrics
4 Identifying Anomalies
According to the discussion in Section 2, the anomalous behaviour of distinct types of
anomalies has different signatures in each of the four main metrics: bandwidth, load,
the number of distinct source sockets and distinct destination sockets. Therefore, we
also apply subspace method to the other 3 parameters of all links to detect anomalous
patterns in each individual traffic feature. The results are put together for
simultaneous analysis as shown in Fig. 2(b). We see that network anomalies exhibit
anomalous behaviour differently in each of the 4 traffic metrics. The detection results
match the initial comment in Table 1. In case of network outage, even though when it
occurs, most of the metrics decrease abnormally but in PCA analysis, such exhibition
is still considered as data variance so it will create spikes in residual vectors. This
makes the detection result of network outage similar to that of FlashCrowds. Besides,
when there is a failure in a part of the topology, network nodes always tend to find
substitutive resource to use, so the variance of traffic metrics is not large enough to
create major spikes.
5 Conclusion
In this paper, PCA subspace method is applied to detect and to identify network
anomalies based on link traffic analysis. With the traffic obtained through simulation
experiments, the efficiency of the proposed method in detecting network anomalies is
proved. We also address the issue of sensitivity of the method that depends on the
rate of traffic sampling. Since PCA is an energy-based analysis, the method is more
accurate when anomaly’s energy (variability) is comparatively high. It is left for
further study to verify the method using traffic from real network environments which
allow us to add parametric substantiality to the proposed method.
References
1. Wikipedia webpage link: http://en.wikipedia.org/wiki/Main_Page
2. Wang, W., Battiti, R.: Identifying Intrusions in Computer Networks Based on Principal
Component Analysis, Technical Report DIT-05-084, University of Trento (2005)
3. Lakhina, A., Crovella, M., Diot, C.: Diagnosing Network-Wide Traffic Anomalies, Proc.
ACM SIGCOMM, Portland, USA (Aug. 2004) 4-6
4. Le, T. Q., Zhanikeev, M., Tanaka, Y.: Component Analysis in Traffic and Detection of
Anomalies, IEICE Technical Report on Telecommunication Management, No.TM2005-66
(March 2006) 61-66
5. Jackson, J. E., Mudholkar, G. S.: Control Procedures for Residuals Associated with
Principal Component Analysis, Technometrics (1979) 341-349
Preventive Congestion Control Mechanisms in ATM
Based MPLS on BcN: Detection and Control Mechanisms
for a Slim Chance of Label Switched Path
Chulsoo Kim, Taewan Kim, Jin hyuk Son, and Sang Ho Ahn
1 Introduction
BcN is the Korean name of NGN that we resolutely called IP based B-ISDN. Most
telecom operators use separate networks for voice and data services with different
protocols and networking technologies. Current packet switched networks (IP routers)
historically have never supported quality of service due to the fact that it only look at
packets and do not keep any state information on individual flows. An individual flow
might be a voice call, video call, file transfer, or web access. The need to integrate
interactive and distribution services, as well as circuit and packet transfer modes into
a universal broadband network are gradually increasing . For supporting QOS on the
user side, network equipment venders try to introduce new mechanisms for
converging circuit and packet switching into one. The nature of IP is random packet
discarding during a congestion status. For provisioning QOS commitments in IP
traffic, ATM based MPLS was introduced as a backbone system for traffic
engineering and smooth migration from legacy networks. On February 28, 2005, there
was a telecommunication disaster in Korea. The main reason for the disaster was call
concentration at a specific time for telephone banking calls.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 505 – 508, 2006.
© Springer-Verlag Berlin Heidelberg 2006
506 C. Kim et al.
PSTN is directed to check the call completion ratio with regard to the SC candidate
number that was entered by an operator.The originating exchange processes the call
attempt, as an incomplete call by sending a busy tone to the user or to the switching
system. However, when judging all of the busy tones to set an SC call should be
because of numerous system errors. This is because the meaning of SC call is that the
call completion ratios regarding the lack of resources in a specific toll/terminating
exchange or that the user is busy. Therefore, it is necessary to measure the statistical
probability of an incomplete call due to a lack of system resources and user busy.
Two congestion control mechanisms were used in PSTN, that is “Automatic call gap”
and “percentage based” congestion control. The time difference between congestion
recognition and proper control action creates improper call/connection restrictions.
The nominal time difference is 5 minutes for most conventional PSTN exchanges.
4 Fuzzy SC Control
Fuzzy logic was first proposed by Lotfi A. Zadeh of the University of California at
Berkeley in a 1965 paper. He elaborated on his ideas in a 1973 paper that introduced
the concept of "linguistic variables", which in this article equates to a variable defined
as a fuzzy set. In telecommunication networks, several research studies which apply
FCS to traffic control have been reported.One of previous research study using
UCR(Uncomplete Call Ratio) with proper fuzzification rules are the same
measurement interval as conventional PSTN exchanges, but they use UCR from
current uncompleted rate at specific interval. However, gathering UCR data takes the
same amount of time as PSTN control mechanisms. The main part of an FCS is a set
of control rules of linguistic form that comprises an expert’s knowledge. In this paper,
we use NCS (Network Congestion Status) and CNCS (Change of NCS) as two inputs,
to create a FCV (Fuzzy Control Value). These scaled inputs NCS and CNCS have
their own fuzzy values in the set {NB (Positive Big), NS (Negative Small), ZO
(Zero), PS (Positive Small), PB (Positive Big)}. The elements of the fuzzy set have
their own triangular shaped membership functions. Using heuristic knowledge to
decide the call blocking rate, we present the next three rules.
If NCS is high, then it increases CNCS fast.
If NCS is not high, but not too low neither, when CNCS is positive, it increases the
call blocking rate fast, but if it is negative or equal to zero, it holds the current call
blocking rate.
If NCS is low, when CNCS is positive, it increases the call blocking rate
moderately, and when CNCS is negative or equal to zero, it decreases call blocking
rate slowly. Overall fuzzy control system in this paper is following
1. Measurement: NCS , CNCS = NCSn − NCSn − 1
2. Fuzzy value NCSF = NCS / α , CNCSF = CNCS / β (Where1 < α , β < 100)
3. Apply Fuzzy rule(MS: Member Ship) (Sncs(f), SCNCS(f))
Sncs ( f ) = {( MS 1, MSvalue ), ( MS 2 , MSvalue )}
Figure 2 explains the effect of call duration vs. admission calls. In multimedia service,
actual call durations are important factor for determining MPLS system capacity.
Fuzzy NCS shows better performance than fuzzy UCR. However, beyond the
engineering capacity quite different results are generated. We generated 0 to 40%
EFCI/EBCI state randomly every specific intervals. In the Figure 3, we assumed that
before the specific measurement interval (not shown in the figure), system was severe
congestion status. The relevant previous UCR indicates 25% call blocking for any
508 C. Kim et al.
reason. The figure3 shows relevant performance for each control methods. FCS
control method that is based on current NCS and CNCS shows better performance.
But in the x-axis near 86 times shows quite different behavior. It is because
congestion status indicator (EFCI/EBCI) has abnormally generated on the time. We
concludes UCS based control methods can not control a burst traffic, especially it can
not control rapid change of call concentration and normal state.
600 2000
500
1500
Offered Call
Number of calls
Adm issible Calls
100
0
0 -3 -2 -1 0 1 2 3 4 5
10 20 30 40 50 60 70 80 Me asuring Time =0
Ca ll Dura tion(m in) Min(inte rva l=5Min)
Fig. 2. Admissible calls vs. Call Duration and Fuzzy UCR vs NCS
5 Conclusions
In this paper, we present preventive congestion control mechanisms for detecting SC
LSP in ATM based MPLS systems. In particular, we introduce a SC LSP detection
method using network signaling information. SC LSP control can handle 208% call
processing and more than 147% success call, than those without control. It can handle
187% BHCA with 100 times less use of exchange memory. We concluded that it
showed fast congestion avoidance mechanism with a lower system load and
maximized the efficiency of the network resources by restricting ineffective machine
attempts.
Scalable DiffServ-over-MPLS Traffic Engineering with
Per-flow Traffic Policing*
1 Introduction
The market for Internet Service Providers (ISP) is positioned to accept new premium
service offerings such as voice over IP (VoIP), video on demand (VoD), and
streaming television. QoS provisioning is essential for correct operation of
abovementioned services.
Integrated services (IntServ) [1] and Differentiated Services (DiffServ) [2],
standardized by IETF, do not meet current requirements to QoS provisioning. The
former scales poorly and the latter can provide QoS only for traffic aggregates.
Support of Differentiated Services [2] in MPLS [3] that was standardized by IETF can
provide QoS guarantees, while keeping network resource utilization at high level, but
also has no microflow policing. Also, Diffserv-over-MPLS [4] requires complex
packet processing, especially on ingress nodes, and most of implementations separate
DiffServ ingress node from MPLS ingress node. One of the performed tasks in
DiffServ-over-MPLS processing is classification, which has either high memory
complexity or high time complexity. Aggregated Bit Vector [5] is one of the
algorithms that can fit requirements and limitations of network processor. Also
DiffServ-over-MPLS TE requires scheduler to support high number of active queues,
*
This research was supported by the MIC, under the ITRC support program supervised by the
IITA.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 509 – 512, 2006.
© Springer-Verlag Berlin Heidelberg 2006
510 D. Siradjev, I. Gurin, and Y.-T. Kim
The proposed scheme is shown in the Fig. 1. All incoming IP packets are classified
into separate flows. Each flow has its own traffic parameters that define Commited
Information Rate (CIR) and Peak Information Rate (PIR) with corresponding burst
sizes. After performing per-flow policing each flow is mapped to certain LSP. MPLS
label stack operations are performed on incoming MPLS packets. As a result, a set of
outgoing LSPs are produced. Each LSP has its assigned bandwidth, which is greater
than or equal to the committed rate of all flows and LSPs that it carries. If there is
exceeding traffic, currently unused bandwidth will be used to send it, or lower priority
traffic will be preempted. The packets that could not be classified into flow will go
through usual IP processing, and MPLS packets with last label popped, will be
processed by classifier as usual incoming IP packet. In case of transit core router,
processing is simpler, since no classification and flow metering is required. If some
flow, requires very fine grained QoS, it can use dedicated LSP, although other flows
of same class-type share the same LSP.
The packet classifier and scheduler, proposed in this paper, have IXA SDK
compatible architecture and can be used in conjunction with other IXA SDK
compatible modules. Both of them consist of 3 major parts: core component (CC)
running on XScale processor, microblock running on microengines, and specific
configuration utility running on XScale processor, which plays the role of interface
between control plane and forwarding plane. Task of the core component is to
configure the classifier operation and manage the rules database while microblock
handles fast packet processing.
The original ABV proposal does not cover the method of search in each of fields
describing only the aggregation technique that exploits the sparseness of bitvectors. In
the proposed range classifier tries are used for search in certain dimension since tries
can provide constant lookup time, at the expense of memory used. In order to
decrease lookup time to some tolerable value, multibit tries, where each node
represents several bits of prefix, are used. To increase update time for rules that have
wildcarded fields, we introduce additional wildcard bitvector. Bitwise OR operation is
performed on it and on the result of search, prior to combining results of individual
search engines.
In order to achieve good scalability in packet scheduler constant computational
complexity related to the number of queues in the system is essential. Bitvector-based
schemes do not have such properties, due to linear complexity of multiplication,
finding first set bit and other operations. In this paper, we propose a scalable low-jitter
DRR scheduler with dual bidirectional linked list for constant complexity. The
solution employed to gain the design objectives is storing the set of active queues in
bidirectional linked list. Using two linked lists and prohibiting sending back-to-back
packets, when there is more than one active queue, allows achieving another goal that
is reducing delay jitter comparing to original DRR. Details of classifier and scheduler
architecture are explained in [7].
3 Evaluations
IXDP2400 development platform was used for implementing proposed DiffServ-
over-MPLS TE architecture. IXDP2400 platform has four 1Gigabit ethernet ports.
512 D. Siradjev, I. Gurin, and Y.-T. Kim
Smartbits 6000C traffic generator was used to generate network load. Tests were done
for different types of processing. This system also has 4 1 Gigabit Ethernet ports. So
this allows us to generate full load for network processor. Number of created rules is
equal to 4,096 in all cases. First test was made with processing that receives IP
packets, classifies them, and performs MPLS label pushing and related stack
operation. According to the results, DiffServ-over-MPLS application with range
classifier can support the line rate of OC-48 when the packet size is 512 bytes. Per-
microblock processing latency measurement shows that the range classifier has the
longest processing time. The detailed discussions of performance analysis and
limitations of IXP2400 Network Processor are shown in [7]. For better performance,
TCAM-based classifier should be used. Implemented scheduler supports 218 queues
and two priority groups.
4 Conclusion
In this paper, we proposed a scalable DiffServ-over-MPLS architecture for Network
Processor with per-flow Policing. Also we proposed range classifier architecture for
Network Processor, based on Aggregated Bit Vector scheme. Application shows
tolerable results in performance test, and the performance can be easily improved by
using TCAM-based classification, instead of software classification. Although the
DiffServ-over-MPLS architecture is well-known, most of its implementations were
designed for hardware, while the implementation on network processor can make it
more flexible. The proposed DiffServ-over-MPLS TE implementation on network
processor should help in the analysis of functional blocks and improving them in
future. Future works include performance enhancement and development of
functional modules of control plane.
References
1. R. Braden et. al., “Integrated Services in the Internet Architecture: an Overview,” RFC
1633, IETF, June 1994.
2. S. Blake et al., “An Architecture of Differentiated Services,” RFC 2475, IETF, December
1998.
3. E. Rosen et al., “Multiprotocol Label Switching Architecture,” RFC 3031, IETF, January
2001.
4. F. Le Faucheur, editor, “Multiprotocol Label Switching (MPLS) support of Differentiated
Services,” RFC 3270, IETF, April 2002.
5. Florin Baboescu and George Varghese, “Scalable Packet Classification,” IEEE/ACM
transactions on networking, vol. 13, No. 1, February 2005.
6. M. Shreedhar, G. Varghese, “Efficient Fair Queuing Using Deficit Round-Robin,”
IEEE/ACM Transactions on Networking, Vol.4, No.3, June 1996.
7. Djakhongir Siradjev, Ivan Gurin, Seung-Hun Yoon and Jeong-Ki Park, “DiffServ-over-
MPLS TE on IXDP2400,” YUANTL-TR-06-NP-DSMPLS-01, July, 2006.
On the Dynamic Management of Information in
Ubiquitous Systems Using Evolvable Software
Components*
1 Introduction
With the ubiquity of emerging ubiquitous devices, our access to data would evolve
exponentially and this continuous evolution in the information and interfaces would
overwhelm the humans. Therefore, we need a mechanism which could evolve both in
terms of information and software.
Different types of software infrastructures for the ubiquitous systems have been
developed in the past years like Gaia [5], Solar System [8] and Context Toolkit [9]
etc. But, they didn’t consider the gradual evolution of the ubiquitous environments
and the composition of ontologies is also done manually. To cope with evolution and
interoperability, it is necessary to separate the overall environment into smaller logi-
cal groups or modules. Also, in these rapidly evolving ubiquitous environments, it
may not even be desirable to build a comprehensive ontology, because the interfaces
of appliances or devices are liable to change. Therefore, we seek to develop onotolo-
gies dynamically. Number of researchers has proposed different dynamic ontology
extraction algorithms for text documents and web pages e.g. Chung Hee Hwang [2],
Giovanni et al [3], and Zhan Cui et al. [4] etc. But, the dynamic extraction of ontolo-
gies in the ubiquitous environment to learn the information about the environment and
user to provide him/her the seamless effect in the interaction with devices and ser-
vices around him was never considered.
*
This wok was supported by MIC and ITRC Project.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 513 – 516, 2006.
© Springer-Verlag Berlin Heidelberg 2006
514 S.S. Murtaza, B. Ahmed, and C.S. Hong
2 System Description
Our system contains two major applications: Server application and User application,
along with several device applications for each home appliance (TV, Bulb etc).
The system employs UPnP [7] protocol and is developed using C# and Intel’s
UPnP SDK. A self explanatory description of the system is shown in figure 1.
On user’s interaction with devices, the user application extracts device name from
the UPnP based device description by extracting the stop words from the manufac-
turer tag, then eliminating stop words1 from the common string of the “Model Name”
and “Model Description” tags of UPnP device description. This is shown in figure 2.
The next step is to find the relations between devices. This is done at the server by
finding the device and sensor relation using association rules [6]. The general tem-
plate of the association rule is given below, where D manifests device and S manifests
sensor. These association rules then become the basis of the device and schedule
classification rules.
D1ŀ D2 ŀ D3 ŀ …..Dn Æ Si (1)
Device Classification Rules:
Let X={X1, X2….. Xn} be the devices and Xi, Xj X where i,j=1 to n
1) If X1, X2 devices affect some common sensors and other uncommon sensors
then they can be regarded as similar devices.
2) Devices always affecting same common sensors are same devices.
3) Same Device: If X set of devices is the ‘Same Device’ and if Xi has the
request of use and it is not available then use Xj such that Xj value of effect
1
Here, stop words represent those words which are not the part of actual name.
On the Dynamic Management of Information in Ubiquitous Systems 515
3 Evaluation
We considered a single room, a user and designed simple UPnP home appliances e.g.
dimmable light bulb, medium size microwave oven, air conditioner with temperature
control facility and etc. Similarly, we deployed different sensors like illumination sen-
sor, temperature sensor and noise sensor. We performed the experiment for a week and
divided the time durations into different slices according to the schedule rule for this
experiment: Breakfast and wakeup timing, office timing, dinner and relaxation timings
and sleep timing. Initially, the user application extracted the names then the server ap-
plication calculated the relationships according to the association and classification rules
of the previous section. We have shown these ontologies and their results in the form of
a tree in the figure 3.These results allow the system to automate the environment and
provide a seamless effect in a number of ways e.g. When the user gets up in the morning
then according to the schedule rules of the previous sections the system will automati-
cally turn on the Light bulb, but if it is not available then by same device rule it will turn
on the lamp. Similarly, many other scenarios are possible.
516 S.S. Murtaza, B. Ahmed, and C.S. Hong
In the future, we would like to publish these extracted ontologies in the form of
RDF to reuse and share this work with other systems
References
1. Syed Shariyar Murtaza,Choong Seon Hong," An Evolvable Software Architec-ture for
Managing Ubiquitous Systems",Proc. of 8th Asia Pacific Network Operation and Manag-
ment Symposium (APNOMS), September 2005,pp 400-409
2. Chung Hee Hwang, ”Incompletely and Imprecisely Speaking: Using Dynamic Ontologies
for Representing and Retrieving Information”, proc. of the 6th International Workshop on
Knowledge Representation meets Databases, July ,1999.
3. Giovanni Modica, Avigdor Gal, and Hasan M. Jamil, “The Use of Machine-Generated On-
tologies in Dynamic Information Seeking”, Proc. of the 9th In ternational Conference on
Cooperative Information Systems, 2001, 443- 448
4. Zhan Cui, Ernesto Damiani, Marcello Leida, Marco Viviani, “OntoExtractor A Fuzzy-
Based Approach in Clustering Semi-structured Data Sources and Meta data Generation”,
proc. of Knowledge-Based Intelligent Information and Engineering Systems, LNCS, 05
5. A Ranganathan, Roy Campbell: A Middleware for Context -Aware Agents in Ubiquitous
Computing Environments, proc. of ACM/IFIP/USENIX Middleware Conference,June,03
6. Han, Jiawei. Kamber, Micheline.: Data Mining: Concepts and Techniques, Morgan Kauf
mann Publishers (2001): ISBN: 1558604898.
7. UPnP forum, http://ww.upnp.org
8. Guanling Chen and David Kotz.: Solar: An Open Platform for Context –Aware Mobile
Applications, proc of the First International Conference on Pervasive Computing (Perva-
sive 2002), Switzerland, June, 2002.
9. Context Toolkit project http://www.cs.berkeley.edu/~dey/context.html
A Shared-Memory Packet Buffer Management
in a Network Interface Card
1 Introduction
A shared-memory packet buffer in a network interface card (NIC) is a large shared
dual-ported memory [4]. Packets for each application are multiplexed into a single
stream. In an intelligent NIC, the packet buffer management algorithm determines
whether to accept or reject each packet. The accepted packet is placed into a logical
FIFO queue; each application has its own queue in a packet buffer [1]. The accepted
packet remains in a buffer until the application retrieves it from the buffer. These
accumulated packets in the buffer can reduce the available buffer space for a next
incoming packet. Once the buffer is full, further incoming packets will be dropped.
Therefore, it is important to reduce packet loss ratio to support any end-to-end
application in a computer network [4]. Buffer management algorithms in a NIC
determine how the buffer space is distributed among different applications. The
design of a buffer management algorithm needs to consider the following two factors
[1]: 1) Packet loss ratio and 2) Hardware complexity. We propose an efficient buffer
management algorithm called Evenly Based Dynamic algorithm (EBDA); EBDA
provides fairness to all the applications while reducing the packet loss ratio.
Tomas Henriksson, et al. [1] proposed protocol processor architecture to offload
the host processor for a high-speed network. The new packet reception, move the
layer 3 and layer 4 processing to an intelligent NIC. The main goal of the protocol
processor is to handle the TCP/IP or the UDP/IP processing at a wire speed.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 517 – 520, 2006.
© Springer-Verlag Berlin Heidelberg 2006
518 A. Uppal and Y. Chu
3 EBDA
DADT reduces the overall packet loss ratio by giving less threshold value to the
applications with larger packet sizes. This results in an increase in packet losses for
applications with larger packet sizes, thus resulting in reducing fairness for
applications with large packet sizes. Therefore, we proposed EBDA that will take
fairness among applications and packet sizes of applications into consideration while
allocating buffer space to each application.
Fig. 1 shows the flowchart of EBDA. In EBDA, the threshold value for an
application ‘i’ with packet size ‘psize(i)’ less than the average packet size
(Σpsize(i)/n) is calculated as shown in equation 3; where n is the number of total
applications. For an application with a packet size greater than the average packet
size, the threshold value is calculated as shown in equation 4.
Our simulation results have shown that by taking packet size factor in the
summation as in equation 3 and 4, instead of multiplication for determining the
A Shared-Memory Packet Buffer Management in a Network Interface Card 519
Y N
α1.(M -Q(t))+ Ȗ1* psize(i ) >= Q(i) α2.(M -Q(t))+ Ȗ2* psize(i ) >= Q(i)
N Y Y N
threshold value for the application, we can reduce the overall packet loss ratio as well
as distribute the packet losses more evenly among the different applications.
T(t)=α1* (M-Q(t))+ 1* psize(i) (3)
We implemented a traffic mix with the average network traffic loads according to [2].
We have taken six applications for our simulations with packet sizes as 8,2,8,1,4,16
Bytes respectively. We have taken a buffer size of 600 packets.
For DA, optimum value of alpha for minimum packet loss ratio for average traffic
load comes out to be 16 and for DADT, optimum value of alpha for different
applications comes out to be 16,14,16,14,16,8 respectively. For EBDA, optimum
values of alpha1, gamma1, alpha2, and gamma2 come out to be 16, 4, 64, and 64
respectively.
Fig. 2 shows the performance of the three algorithms (EBDA, DA, and DADT)
for different loads. Load has been varied from 0.5 to 0.9. As seen in Fig. 2, EBDA
has the least packet loss ratio for all of loads. Notice that the performance
difference increases more at higher loads. As the load is increased, applications
with larger packet size tend to increase their queue length to values greater than
their threshold values frequently. Since, EBDA utilizes the buffer space more
efficiently, providing fairness to all the applications; EBDA can reduce the packet
loss ratio significantly.
520 A. Uppal and Y. Chu
DA
0.15
0.05
0
0.5 0.6 0.7 0.8 0.9
Load
Fig. 2. Packet loss ratio vs. Load for EBDA, DADT, DA for the average traffic load
5 Conclusions
Three buffer management algorithms are implemented for our simulations: 1)
Dynamic algorithm (DA); 2) Dynamic Algorithm with Dynamic Threshold (DADT);
and 3) Evenly Based Dynamic algorithm (EBDA). EBDA provides more fairness to
all the applications and utilizes buffer space efficiently, which makes it different from
DA and DADT. For the traffic mix with the average network traffic loads, the EBDA
improves the packet loss ratio by 16.7% as compared with DA, and by 11.8% as
compared with DADT.
References
1. T. Henriksson, U. Nordqvist, D. Liu, Embedded Protocol Processor for fast and efficient
packet reception, IEEE Proceedings on Computer Design: VLSI in Computers and
Processors, vol. 2, pp. 414-419, September 2002.
2. U. Nordqvist, D. Liu, Power optimized packet buffering in a protocol processor,
Proceedings of the 2003 10th IEEE International Conference on Electronics, Circuits and
Systems, vol. 3, pp. 1026-1029, December 2003.
3. M. Arpaci, J.A. Copeland, Buffer Management for Shared Memory ATM Switches, IEEE
Communication Surveys, First Quarter 2000.
4. F. Kamoun, L. Kleinrock, Analysis of Shared Finite Storage in a Computer Network Node
Environment under General Traffic Conditions, IEEE Transactions on Communications,
vol., COM-28, pp. 992-1003, July 1980.
5. A. K. Choudhury, E.L. Hahne, Dynamic Queue Length Thresholds for Shared-Memory
Packet Switches, IEEE/ACM Transactions on Communications, vol. 6, no. 2, pp. 130-140,
April 1998.
6. Yul Chu, Vinod Rajan, An Enhanced Dynamic Packet Buffer Management, In the
proceedings of the 10th IEEE Symposium on Computers and Communications (ISCC'05),
Cartagena, Spain, June 2005.
An Adaptive Online Network Management Algorithm for
QoS Sensitive Multimedia Services
Abstract. The explosive growth of new multimedia services over the Internet
necessitates efficient network management. Improved network management
systems are expected to simultaneously provide diverse multimedia traffic ser-
vices and enhance network performance. In this paper, we propose a new online
network management algorithm that implements adaptation, reservation, and
call-admission strategies. Our online approach to network management exhibits
dynamic adaptability, flexibility, and responsiveness to the current traffic condi-
tions in multimedia networks.
1 Introduction
Multimedia is a keyword in the evolving information age of the 21st century. In re-
cent years, the growth of multimedia applications that can be represented through
audio and, video streams, images and animations has greatly increased the research
interest in quality of service (QoS). The ultimate goal of network QoS support is to
provide applications with a high-quality data delivery service [1]-[2].
The different types of multimedia service provided over networks not only require
different amounts of bandwidth but also have different policy assumptions. The het-
erogeneous multimedia data usually categorized into two classes according to the
required QoS: class I (real-time) and class II (not real-time). The class I data type has
a higher priority than class II data type and so a multimedia network should take into
account the prioritization among different multimedia traffic services [1]-[2].
Internet multimedia services have continued to emerge thanks to the benefits asso-
ciated with the stateless architecture of the Internet Protocol (IP), and has made the
provision of QoS-sensitive multimedia data services an area of great importance.
However, the Internet is currently not designed to support the performance guarantees
such as bounded delay and minimum throughput that are generally required for the
higher priority class I applications [3].
QoS provisioning requires the management of admission control [1]-[3]. An essen-
tial role of call admission is to protect admission-controlled traffic from non-
admission-controlled traffic. The bandwidth of a link on the Internet is shared
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 521 – 524, 2006.
© Springer-Verlag Berlin Heidelberg 2006
522 S. Kim and S. Kim
dynamically between class I and class II data services. Since - each service has differ-
ent operational requirements - different admission control rules are applied to each
application type. For example, based on traffic priority, there can be class I applica-
tions with strict admission control rules and class II applications with non-controlled
admission rules [3].
The call-admission mechanism is used to reserve bandwidth. In a reservation pro-
cedure, some of the available bandwidth is reserved for use by higher priority traffic
services. Therefore, admission-controlled class I data services can actually benefit
from bandwidth reservations for QoS guarantees [1]-[3].
Efficient network management requirements control decisions that are dynamically
adjustable. However, at any point in time the future rate of traffic arrival is generally
not known, and there can be dramatic short-term variations in traffic patterns. These
control decisions therefore have to be made in real time.
Online algorithms [4] are natural candidates for the design of efficient control
schemes in QoS-sensitive multimedia networks. An algorithm employing online com-
putations is called an online algorithm, and the term ‘online computation problem’
refers to decision problems where decisions must be made in real time based on past
events without information about the future.
Motivated by the above discussion, we propose a new online network management
algorithm for QoS-sensitive multimedia networks. Our algorithm is designed to han-
dle control decisions in an online manner. Due to the uncertain network environment,
an online strategy based on real time measurements of the current network conditions
plays an important role in determining the network performance.
The important features of our algorithm are (i) the inclusion of a QoS guarantee
that does not reduce the network capacity, (ii) the ability to adaptively control conges-
tion so as to maximize network performance, (iii) a low complexity that makes it
practical for real-world implementation, and (iv) the ability to respond to current
network traffic conditions for appropriately balancing the performance between con-
tradictory QoS requirements.
where N i and Bi are the number of class I data requests and the corresponding band-
widths of data type i, respectively. Therefore, by using this traffic window, we can
adjust the amount of the reserved bandwidth (ResB) at every unit_time, which is more
responsive to changes in the network condition after the bandwidth has been reserved.
A BB monitors the traffic load of each link in its own domain and updates the link
database by calculating the current amount of unoccupied (UBlink) and reserved
(RBlink) bandwidth. The total available bandwidth (ABlink) for class I services can be
estimated as the sum of UBlink and RBlink:
References
1. Sungwook Kim and Pramod K. Varshney, “An Adaptive Bandwidth Allocation Algorithm
for QoS guaranteed Multimedia Networks ”, Computer Communications 28, pp.1959-1969,
October, 2005.
2. Sungwook Kim and Pramod K. Varshney, “An Integrated Adaptive Bandwidth Manage-
ment Framework for QoS sensitive Multimedia Cellular Networks”, IEEE Transaction on
Vehicular Technology, pp.835- 846, May, 2004.
3. J. Lakkakorpi, O. Strandberg and J. Salonen, "Adaptive Connection Admission Control for
Differentiated Services Access Networks," IEEE Journal on Selected Areas in Communica-
tions, Vol. 23, No. 10, October 2005, pp. 1963-1972.
4. Yossi Azar, Online Algorithms - The State of the Art, Springer, 1998.
Improved Handoff Performance Based on
Pre-binding Update in HMIPv6
1 Introduction
Mobile IPv6 [1,2] handoff incurs high handoff latency, data loss, and global
signaling. Basically, Fast Mobile IPv6 (FMIPv6) [3] reduces handoff latency by
link layer (L2) triggers and prevents data loss by creating a bi-directional tunnel
between a mobile node’s previous subnet’s access router (oAR) and next subnet’s
access router (nAR). HMIPv6 [4] prevents global handoff signaling by appointing
a MAP that acts like a local Home Agent (HA). In Mobile IPv6 and HMIPv6,
no information is exchanged among ARs. Therefore, only after completing L2
handoff, an MN can receive information regarding the AR, to which the MN will
handoff via an agent advertisement message. On-going communication sessions
with other hosts are impossible before completion of this handoff process, which
is the major portion of overall handoff latency [5,6].
In the proposed mechanism, an AR can learn information regarding its geo-
graphically adjacent ARs - typically, global address, L2 identifier, and the prefix
information of ARs that are currently being advertised. The current AR that the
MN is visiting would be able to inform the MN of the prefix information of ARs
to which the MN would likely handoff. After completion of the Address Auto-
configuration (AA) process, the MN transmits an incomplete binding update
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 525–529, 2006.
c Springer-Verlag Berlin Heidelberg 2006
526 J. Jeong, M.Y. Chung, and H. Choo
message to a MAP, and then the MAP performs a Duplicate Address Detec-
tion (DAD) process using this message. Through this signaling flow, there is a
remarkable decrease in handoff latency.
2 Related Works
This section provides a brief overview of the differences to be taken into account
for the various approaches to reduce the handoff latency. Basic MIPv6 [1,2],
Anticipated FMIPv6 [3], Hierarchical Mobile IPv6 [4], Hierarchical Mobile IPv6
with Fast handover (F-HMIPv6) [3,8], FMIPv6 for HMIPv6 (FF-HMIPv6) [9].
registration MAP’s RCoA) to MAP3, and then MAP3 transmits two messages
to MAP1. In this case, MAP1 deletes MAP2’s RCoA in the MN’s list since
it contains the MAP’s RCoA and records MAP3’s RCoA. MAP1 and MAP3
are then linked through the registration. Therefore, the proposed scheme is not
required to send the binding information to the HA and CNs. Fig. 1 represent
the signal flows of the proposed handoff process, and only the Inter-MAP domain
handoff is described in detail.
4 Performance Evaluation
First, the handoff latency is studied - the sum of the L2 and L3 handoff la-
tency. The handoff latency of basic Mobile IP is defined as the period between
the disconnection of the MN’s wireless link and reception of AR’s binding ac-
knowledgement by MN. It is used for Intra-MAP domain handoff. In Inter-MAP
domain handoff, handoff latency is the time from when MN triggers link-down
in the current network to when the MN receives HA’s first binding acknowl-
edgement after handoff. The proposed scheme presents the minimum latency.
For Inter-MAP domain handoff, HMIPv6 presents the largest handoff latency.
Note that the Inter-MAP domain handoff improvement of the proposed scheme
is very large. This is due to the fact that the proposed scheme performs the AA
and DAD processes in advance. MIPv6 presents many packet losses due to the
large handoff latency, and HMIPv6 presents decreased packet losses than the
base MIPv6 through the advantages of the MAP entity. The proposed scheme
presents superior performance without packet loss. In the proposed scheme, the
528 J. Jeong, M.Y. Chung, and H. Choo
MAPs buffer and forward packets during the handoff period, this improves over-
all performance.
From analysis [7], the total signaling cost of the packet forwarding scheme is
smaller than the HMIPv6 up until 8 forwarding steps (q≤8). q represents the
maximum number of forwarding link. However when q≥9, the cost of the packet
forwarding scheme becomes greater than the HMIPv6 one, and the MN sends
the registration message to the HA and the CNs, removing all of the previous
links among MAPs. Although at q≥9, the total signaling cost changes little. In
the worst case, the total signaling cost of the proposed scheme is smaller than
those of the HMIPv6 and the Forwarding up until 17 forwarding steps (q≤17).
5 Conclusion
Acknowledgment
References
Soomi Yang
Abstract. For an efficient role based access control using attribute certificate in
highly distributed computing environments, we use a technique of structuring
role specification certificates. The roles are grouped and made them into the re-
lation tree. It can reduce management cost and overhead incurred when chang-
ing the specification of the role. Further we use caching of frequently used role
specification certificate for better performance in case applying the role. And
for the global space reduction, we also consider the issue of tree normalization.
In order to be scalable distribution of the role specification certificate, we use
multicasting packets. In the experimental section, it is shown that our proposed
method is secure and efficient.
1 Introduction
A role based access control using attribute certificates can provide more flexible and
secure environments than that using only public key certificates[1,2]. And the use of
attribute certificates can provide more adaptable scheme by the use of role specification
certificates. Highly distributed environments usually need support of the authorization
of resources at varying levels of access. Furthermore, it needs the interactions of highly
collaborating entities to be secure. However, the distributed environments could not
have any central or global control. Therefore for security of highly distributed environ-
ments, we distribute the role specifications according to the levels of access. It accords
with the characteristics of the distributed environments and sometimes is inevitable. We
distribute the privileges. In addition, we group roles of privileges, which is different
from the typical methods which group subjects only[1,2].
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 530 – 533, 2006.
© Springer-Verlag Berlin Heidelberg 2006
On the Security of Attribute Certificate Structuring 531
The distribution of updated role specification certificates makes use of the multicast
communication. Let F(l) ( 0 ≤ l ≤ h )be the frequency of the transmission of a role
specification certificate s i in order to be successfully delivered to all W(l) receivers.
The probability that one of these W(l) receivers (say w) will not receive the updated
role specification if it is transmitted once is equal to the probability of packet loss, p,
for that receiver. Let Fw be the frequency of role specification transmissions necessary
for receiver w to successfully receive the role specification certificate. Since all the
packet loss events for receiver w, including replicated packet and retransmissions, are
mutually independent, Fw is geometrically distributed. Thus, the average expected
frequency of the role specification packet transmission can be computed as following:
∞ ∞
E[ F (l )] = ¦ P[ F (l ) ≥ f ] = ¦ (1 − (1 − p f −1 )W ( l ) ) (1)
f =1 f =1
We can compute E[F(l)] numerically using Equation (5) by truncating the summation
when the fth value falls below the threshold.
For an application of the role specification certificate, 2*(h-l) packets should be suc-
cessfully transmitted. It forms a path through the role specification tree from a re-
questing node to a node having requested role specification certificate. To improve
the application of the role specification, caching scheme can be adopted. Let G(l) be
the frequency of the transmission of packets for the successful delivery of a role
specification certificate si to a requesting node. If the probability of having cached
role specification certificate is q, the average expected frequency of the role specifica-
tion packet transmission can be computed as following from the similar induction of
role distribution.
∞ ∞
E[G (l )] = (1 − q )¦ P[G (l ) ≥ f ] = (1 − q )¦ (1 − (1 − p f −1 ) 2 ( h − l ) ) (2)
f =1 f =1
A cached role specification does not need any packet transmission. It holds some
space of a requesting device. q can be easily computed through the Markov chain
state transition inspection.
532 S. Yang
For global space reduction, we can normalize the role specification. In normalized
role specification tree, privilege should appear only in one certificate. The number of
certificates in level l is Gnorm C h +1−l = 2 l . Thus the total number of certificates can be
Gnorm
induced to ¦ i =1 Gnorm C i . It can include unused role group. However it generally
significantly reduces the space compared to the space used for naturally evolving and
diminishing role specifications without any regulation. Furthermore it reduces com-
munication cost incurred when changing the specification of the role as shown in
Table 1.
Table 1. The number of packet transmissions for changing the specification of the role
Un-normalized normalized
ungrouped grouped ungrouped grouped
f = 10, p =0.1 2.54 1.39 1.66 1.21
f = 200, p = 0.9 45.53 20.27 26.30 14.74
However normalization increases path length and incurs more packet transmission
when the subject is going to apply the role specifications. Therefore the normalization
should be adopted with caching conforming to the characteristics of individual
environment.
3 Performance Evaluation
We measure the expected number of packet transmission, E[F(l)] and E[G(l)], for the
performance comparison. For each given caching ratio q, we can inspect the effects to
the average packet transmission. The ‘cached’ case of Fig. 1 shows the frequency
grouped
cached
45
40
35
30
25
E
20
15
10
0
0.5
0.4 1
0.8
0.3
0.6
0.2 0.4
0.2
q 0.1 0
p
variations by packet loss p and caching ratio q with f=100. When the packet loss is
small, the difference is small. However, as the packet loss gets bigger, it suffers more
increasing packet transmission. Fig. 1 shows the plot of the expected packet transmis-
sion E[F(l)] and E[G(l)] for packet loss p and the caching ratio q. Fig. 1 shows the
greater increase in E[F(l)] where the roles are grouped than in E[G(l)] where the roles
are grouped and the role specifications are cached. If we take a specific sample case,
from the values given in Fig. 1 we can see that total number of the packet transmis-
sion should be greatly decreased when the role specifications are cached. However if
the nodes are distributed geographically and the packet loss is more often, the per-
formance enhancements gained when the role specifications should be changed are
overwhelming.
4 Conclusion
As an efficient access control using attribute certificate, we use the technique of struc-
turing role specification certificates and reinforce it through caching them. It can
reduce the management cost and overhead incurred when changing and applying the
specification of the role. We grouped roles, made the role group relation tree, and
showed the model description. It provides the secure and efficient role updating, ap-
plying and the distribution. For scalable role specification certificate distribution, we
used multicasting packets. The performance enhancements are quantified with taking
into account the packet loss.
References
1. ITI (Information Technology Industry Council), Role Based Access Control ITU/T. Rec-
ommendation X.509 | ISO/IEC 9594-8, Information Technology Open Systems Interconnec-
tion-The Directory: Public-Key and Attribute Certificate Frameworks (2003)
2. S. Farrell and R. Housley, An Internet Attribute Certificate Profile for Authorization, IETF
RFC 3281, (2002)
3. Sandro Rafaeli, David Hutchison, A Survey of Key Management for Secure Group Com-
munication, ACM Computing Surveys, Vol. 35, No. 3 (2003)
Performance Analysis of Single Rate Two Level Traffic
Conditioner for VoIP Service
Dae Ho Kim, Ki Jong Koo, Tae Gyu Kang, and Do Young Kim
Abstract. In this paper, for the combination of DiffServ and RSVP, we propose
traffic conditioning algorithm. Through the proposal, we expect the service
quality improvement of real time application, especially voice traffic, by sepa-
ration of other data traffic.
1 Introduction
In this section, we propose Single Rate Two Level Traffic Conditioner which meters
and shapes input traffic according to its reservation style in EF class of DiffServ in-
gress node. Single Rate Two Level Traffic Conditioner has two leaky buckets which
one is for RSVP Fixed-Filter style traffic and the other for other RSVP reservation
style traffic and just admission controlled traffic.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 534 – 537, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Performance Analysis of Single Rate Two Level Traffic Conditioner for VoIP Service 535
Fig. 1. Single Rate Two Level Traffic Conditioner architecture in the Ingress node of DiffServ
network
Single Rate Two level Traffic Conditioner consists of two leaky buckets. We call
these two leaky bucket, master leaky bucket and slave leaky bucket, because slave
leaky bucket rate is controlled by master leaky bucket rate. Master Leaky Bucket
(MLB) is reserved by RSVP Fixed-Filter (FF) style traffic. Its initial leaky bucket rate
Rmi is 0 and is increased by RSVP FF style reservation (1). Slave Leaky Bucket
(SLB) is same with general leaky bucket in DiffServ ingress node and used by RSVP
WF and SE style reservation and just admission controlled traffic. Its initial leaky
bucket rate Rsi is EF class maximum rate RMAX (1). RMAX is the maximum amount of
in-profile traffic in order for a network administrator to guarantee a high forwarding
probability to that traffic.
Rmi = 0 = Rm , Rsi = RMAX = Rs. (1)
All these mechanisms are performed dynamically by RSVP soft state characteristics,
that is, if destination indicates service quality degradation to source, source will
536 D.H. Kim et al.
request service quality guarantee to network elements with RSVP refresh message.
Ingress node receiving RSVP Path message for refresh will update service parame-
ters, bucket rate and size. Fig. 2 shows system architecture of DiffServ ingress router
when our proposal is used.
3 Simulation Result
Fig. 3 shows network delay and delay variance of voice traffic of each class in Diff-
Serv network. According to classification and priority scheduling of packet, high class
packets get higher quality of service. But delay of EF class is higher when offered
load of EF class increases more than 90% of limited in-profile traffic. This is effect of
leaky bucket shaper. Fig. 3 shows also poor quality of EF class in delay variance
aspect. In this result, we can know also importance of admission control of EF class
and dynamic control.
Fig. 4 shows network delay and delay variance of voice traffic of each class in RSVP-
enabled DiffServ network. Network delay of RSVP uncontrolled classes, EF, AF and
BE, show same pattern with network delay in DiffServ network. But RSVP controlled
class (RSVP EF in Fig. 4) shows high performance of low delay and delay variance.
4 Conclusion
Through this thesis and simulation, we can know importance of traffic admission
control and dynamic rate control of shaper in DiffServ network. Decreasing effort of
delay and delay variance in shaper of DiffServ ingress node must be studied and ana-
lyzed continuously. It could be obtained by RSVP mapping to DiffServ network, we
think. So more detail study and research are needed.
References
1. R. Braden, L. Zhang, S. Berson, S. Herzog, S. Jamin, “Resource Reservation Protocol
(RSVP) – version 1 Functional Specification,” RFC 2205, Sep 1997.
2. S. Blake, D. Black, M. Carson, E. Davies, Z. Wang, W. Weiss, “An Architecture for Differ-
entiated Services,’ RFC 2475, Dec 1998.
3. R. Balmer, F. Baumgarter, T. Braun, M. Gunter, “A concept for RSVP over DiffServ,”
Computer Communications and Networks, Ninth International Conference on, p.p. 412 -
417, 2000.
An Architectural Framework for Network Convergence
Through Application Level Presence Signaling
Atanu Mukherjee
Chief Architect
Cognizant Technology Solutions Corporation
Tel.: 973.368.9300
[email protected]
Abstract. Over the past few years we have witnessed steady progress toward
convergence in communications. We believe we are now witnessing
convergence beyond just media transport. Further, abstract application level
signaling technologies are poised for takeoff and will serve as the unifying
fabric for multi-dimensional convergence. Our research shows that such a
signaling mechanism known as “presence”, combined with advances in
namespace and security technology, will enable ultimate service innovations.
Keywords: presence management, SIP, application layer signaling, network
convergence.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 538 – 541, 2006.
© Springer-Verlag Berlin Heidelberg 2006
An Architectural Framework for Network Convergence 539
3 Conclusion
References
1. http://www.verisign.com
2. http://www.illuminet.com
3. http://www.netnumber.com
4. “Verisign Plans to Pay $1.2 B in stock for Illuminet Holdings”, Wall Street Journal, 26th
September, 2001
5. Private Escrow Key Management : A Method and Its Issues, Edward M Scheidt, TECSEC,
Key Escrow Issues Meeting, NIST, Gaithersburg, Maryland, Sept., 6, 1995
Security Approaches for Cluster Interconnection in a
Wireless Sensor Network
1 Introduction
A wireless sensor network is characterized by a collection of nodes, which do not
utilize a network infrastructure. The low processing capability, low memory
availability, and the battery-based energy consumption are also limiting factors that
should be taken into consideration. In this sensor network environment, due to the
limitation on the transmission range, in order for two nodes, A and B, to
communicate, they must utilize multi-hop routing over intermediate nodes between
them. In this sense, nodes act not only as hosts, but also as routers, receiving and
transmitting messages whose end-destinations are other nodes.
Because of the nodes limitations, the utilization of a pure public-key infrastructure
(PKI) becomes unviable because of the processing time and consequently because of
the energy consumption necessary for the encryption and decryption of messages.
One way to reduce the processing time and the energy consumption in message
passing is the utilization of symmetric keys, both in routing and data exchange. A
sensor network can be divided in clusters, in the attempt to localize group sensors by
the characteristics of the services they use and provide, with the goal of optimizing
performance [1].
In this article we present a hybrid secured protocol which guarantees group key
establishment, route discovery and secure message passing. The feasibility of this
protocol will be proved by a simulation. We simulate an environment of a variable
number of clusters conclude that a significant increase in the number of nodes in the
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 542 – 545, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Security Approaches for Cluster Interconnection in a Wireless Sensor Network 543
network does not change the performance of the interconnection between any two
clusters. Our discussion focus in section II we will present the proposed protocol we
will also present the transparent interconnection of clusters. We present in section III
the previous simulations and the results obtained, and in section IV we conclude the
article.
CH1 → CH2: (type, IDch2, IDch1, IDsns, KUsns, IDsns1, Signature sns, (3)
Signature ch1)
CH2 → CH1: (type, IDch1, IDch2, IDsns, Signature ch2); with Signature
(4)
ch2 = Sign KRch2(HASH(type, IDsns, IDch2, IDch1, KUch1, IDsns1))
CH1 → Sensor: (type, IDsns, IDch2, IDch1, KUch1, IDsns1, Signature ch2) (5)
Sensor → CH: (tipo, IDch, IDsns, Signature gr); with Signature gr =
(6)
EKs(Hash)
544 A. Gava Menezes and C. Becker Westphall
Where (1) is the request to join the cluster and obtain the group key. (2) is the response
to the join request and transmission of the group key. (3) is the requests forward to join
the cluster to the destination cluster head. (4) is the configuration response of the sensor
to accept the new cluster head. (5) is the reconfiguration message of the sensor to belong
to the new cluster head. (6) is the request to obtain a new group key.
In this article, we are focused in avoiding the processing overload and energy
consumption of the sensors. The idea is to transfer the key agreements to the cluster
heads. This way, the processing time to establish a session key will be lesser if it is
established for any two entire clusters instead of being established for just two
sensors. If a sensor wants to communicate with another sensor inside the cluster, it
simply does it by using the group key. When the destination sensor is in another
cluster, the entire message path will be done using different session keys. First of all,
the originator sensor encrypts the data messages using its own cluster group key. The
cluster head uses the session key that was established with the destination cluster
head. The destination cluster head uses its own group key. In the last step, the
destination sensor accesses the secured information with the group key.
One of the main advantages of this approach is achieved when any other sensors of
any of these two clusters want to communicate. The entire route discovery will be
optimized due to the fact that the most complex process has already been done.
This way, we have a transparent cluster interconnection, because the sensor doesn’t
need to know in what cluster is the sensor that it wants to communicate to. This
approach will be the base of our simulations to demonstrate that by increasing the
number of the sensors and consequently the number of clusters in the network, the
messages exchange performance will be the same, having only the increasing of
the propagation time of the message through the neighboring nodes.
3 Simulations
The simulations had their main focus on the proposed protocol performance for the
cluster interconnection, being necessary the creation of an environment in the
Network Simulator-2 (NS-2). As the presented solution must be implemented in the
routing layer, the Ad-hoc On demand Distance Vector (AODV)[2] protocol was used,
with the SAODV (Secure AODV)[3] security requirements.
In our simulations, we simulate the 512 bytes CBR (Constant Bit Rate) messages
being sent through any two sensors of any two different clusters, having each one 250
sensors and these sensors were distributed evenly inside the cluster. In this work we
did not consider the mobility of the sensors. We considered frequencies beyond
8MHz, like the 16MHz, 32MHz and 64MHz frequencies.
The previous simulations proved the expected performance of the proposed protocol.
In the Fig. 1, all four curves have a small angular coefficient, despite the fact that they
seem have a bigger inclination curves from the tenth cluster. In the collected values, we
can notice a gradual increase of the time due to the increasing number of clusters.
Increasing the number of clusters, the time for the message to reach the destination
also increases, because the distance increases in the same proportion as the number of
clusters. The propagation time is more relevant than the time to calculate the hash
function.
Security Approaches for Cluster Interconnection in a Wireless Sensor Network 545
Fig. 1. The average time to send a CBR message of 512 bytes over the number of clusters in the
network without any pre-established route
4 Conclusions
The present work tried to adopt the AODV protocol to a clustered sensors network
environment. A key establishment and a mechanism to interconnect securely any two
clusters was proposed. Analyzing the simulations results, we can notice an even
performance in the clusters interconnection, even when the number of clusters was
considerably increased. Some disadvantages of this approach must be exposed. The
way the messages pass through can overload the cluster head, becoming a bottleneck
to the extra cluster communication. Another point that can be a disadvantage is the
way that the message authentication between intermediate cluster heads are treated,
where there can be a delay to identify an anomalous message. However, we do
believe that the benefits pointed out in this work are bigger than those casual
disadvantages. The main attractive point of this approach is to have a constant
performance in a message exchange between any two clusters, even with the
increasing of the number of clusters in the network.
References
1. Bechler, M., Hof, H. J., Kraft, D., Pählke, F. e Wolf, L. (2004) “A cluster-based security
architecture for ad hoc networks”. IEEE INFOCOM.
2. Perkins, C. e Belding-Royer, E. (2003) “Ad hoc on-demand distance vector (AODV)
routing”. IETF Request for Comments, RFC 3561. July.
3. Zapata, M. G., e Asokan, N. (2002) “Securing ad hoc routing protocols”. Proc. ACM
Workshop on Wireless Security (WiSe), ACM Press, pp. 1-10.
4. Kamal, A. B. M. (2004) “Adaptive Secure Routing in Ad Hoc Mobile Network”. Master of
Science Thesis. Royal Institute of Technology. Sweden. November.
A Resource-Optimal Key Pre-distribution Scheme with
Enhanced Security for Wireless Sensor Networks
Tran Thanh Dai, Al-Sakib Khan Pathan, and Choong Seon Hong*
1 Introduction
In this paper, we present a new key pre-distribution scheme utilizing the advantages
of two existing schemes [1] and [2]. Our analysis shows that our combined scheme
performs better than each of these two schemes and ensures enhanced security than
most of the other existing schemes. Principal contributions of our work are: (1) Con-
siderable improvement in sensors’ resource usage while keeping security as the top
priority (2) Rigorous guarantee of successfully deriving pairwise keys that enable
node-to-node authentication and (3) Better network resilience as compromising one or
more pairwise keys does not influence the remaining pairwise keys
* This work was supported by the MIC and ITRC projects. Dr. C. S. Hong is the corresponding
author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 546 – 549, 2006.
© Springer-Verlag Berlin Heidelberg 2006
A Resource-Optimal Key Pre-distribution Scheme 547
ith column of U, Uc(i) always go together when assigned to a sensor node. Now, let us
explain how a common key could be found between two nodes. Assume that sensor Si
and sensor Sj contains [Lr(i), Uc(i)] and [Lr(j), Uc(j)] respectively. When Si and Sj need
to find a common secret key between them for communication, they first exchange
their columns, and then compute vector products as: Si: Lr(i) × Uc(j) = Mij and Sj:
Lr(j) × Uc(i) = Mji . As M is the symmetric matrix, definitely, Mij = Mji . Mij (or Mji) is
then used as a common key between Si and Sj.
Modified Blom’s Symmetric Key Generation Scheme [3]. In this scheme, as long
as no more than λ nodes are compromised, the network is perfectly secure (this is
referred to as the λ -secure property). Increasing λ results in greater network resil-
ience but also results in higher memory usage within each sensor node. During the
pre-deployment phase, a (λ + 1) × N matrix (where N is the maximum number of
sensor nodes in the network) G over a finite field GF(q) and λ (the security parame-
ter discussed earlier) are constructed. G is considered as public information; any sen-
sor can know the contents of G, and even adversaries are allowed to know G. In order
to achieve the λ -secure property any λ + 1 columns of G must be linearly independ-
ent. Let p be a primitive element of GF(q) and N < q. Then, each nonzero element in
GF(q) can be represented by some power of p, namely pi for some 0 < i ≤ q − 1 . A
feasible G can be designed as follows:
ª1 1 1 " 1 º
«p p2 p3 " p N »»
«
« p2 ( p 2 )2 ( p3 )2 " ( p N )2 »
« »
« # »
«¬ p λ ( p 2 )λ 3 λ N λ»
( p ) " ( p ) ¼ (λ +1)× N
K located in the ith row and jth column. In practice, G can be created by the primitive
p of GF(q). Therefore, when storing the kth column of G at node k, it is only neces-
sary to store the seed pk at this node, any node can generate the column given the
seed. After deployment, two sensor nodes i and j can find the pairwise key between
them by exchanging their columns of G and using their private rows of matrix A to
compute, K ij = K ji = A(i ).G ( j ) = A( j ).G (i ) where A(i) and G(j) represent the
ith row of A and jth column of G respectively.
As we stated earlier that, our scheme is a combination of the two above-mentioned
schemes with significant modifications so that it could be more apposite for the mem-
ory-constrained trait of wireless sensor networks. In our scheme, we only store the
keying information in the sensor nodes, which are eventually used to derive two key
halves to constitute a secret pairwise key. When two nodes want to communicate and
are within the communication ranges of each other, the first key half is generated by
548 T.T. Dai, A.-S.K. Pathan, and C.S. Hong
the LU key decomposition scheme while the second half is generated by modified
Blom’s key computation scheme. The pairwise key used to secure the communication
link between two nodes is derived based on these two halves, i.e., a concatenation of
the two halves or using one way hash functions (Figure 1).
Up to this step, both Si and Sj have the two halves of the pairwise key. To derive
the pairwise key, the simplest way is to concatenate the first half with the second half.
The pairwise key can also be created by using a one-way hash function. This key is
stored in the memory of the two sensors for the rest of their communication. In our
scheme, to reduce energy overhead for encryption and decryption of information
exchanged between a pair of non-neighboring sensor nodes, each sensor node has a
message relaying function. That is, messages exchanged between two non-
neighboring nodes are encrypted and decrypted only by the concerned nodes using a
common pairwise key established as described in the steps. The intermediate nodes
only have to relay the messages to the receiving node. They do not need to understand
the contents of the messages. Hence, they do not need to encrypt or decrypt the mes-
sages, which saves the computation and energy power.
Fig. 2. Network-wide Memory usage in our scheme and [1] (left) if λ is 50% of the total no. of
nodes (right) if λ is 25% of the total no. of nodes [λ could be smaller]
References
[1] Choi, S. and Youn, H., “An Efficient Key Pre-distribution Scheme for Secure Distributed
Sensor Networks”, EUC Workshops 2005, LNCS 3823 (2005) 1088-1097.
[2] Blom. R., “An optimal class of symmetric key generation systems”, Advances in Cryptol-
ogy: Proceedings of EUROCRYPT 84 (Thomas Beth, Norbert Cot, and Ingemar Ingemars-
son, eds.), Lecture Notes in Computer Science, Springer-Verlag, 209 (1985) 335-338.
[3] Du, W., Deng, J., Han, Y. S., Varshney, P. K., Katz, J., and Khalili, A., "A Pairwise Key
Predistribution Scheme for Wireless Sensor Networks", ACM Transactions on Information
and System Security, Vol. 8, No. 2, May (2005) 228-258.
Intelligent Home Network Service Management Platform
Design Based on OSGi Framework
1 Introduction
Along with the explosive growth of broadband access and Internet services, net-
worked home appliances and devices are also increasing, and the expectations on
various home network services are growing. Because of this situation, the SP (Service
Provider) is now conceptualizing the Blue Ocean that makes the competition irrele-
vant by offering various home network services and fulfilling the customer’s expecta-
tions of a better quality of life through intelligent home network services.
Fig. 1. The infrastructure and base technologies for the intelligent home network service [1].
The elements of home network technology described in the figure are the home gateway, home
gateway middleware and home appliances access technologies.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 550 – 553, 2006.
© Springer-Verlag Berlin Heidelberg 2006
IHSM Platform Design Based on OSGi Framework 551
In the past 3-5 years, various researches based on the OSGi Framework have been
done[2][3][4][5]. Haitao Zhang presented a mobile-agent and OSGi-based three-tier
control sys-tem architecture for a smart home [6]. Akihiro Tsutsui suggested a Man-
agement Ar-chitecture and Distribution Framework for Home Network Services at an
NGN work-shop in 2005 [7]. Daqing ZHANG presented an OSGi-Based Service
Infrastructure for Context Aware Automotive Telematics [8]. Xie Li proposed an
OSGi-based home network system [9].
In this paper, we propose an open standard-based, intelligent service management
platform that aims to provide a flexible and extensible platform for building intelli-
gent home network services. To have greater flexibility and efficiency in managing
home network services, we further propose an AAA proxy, service provisioning,
security agent, sip proxy service, and others.
The rest of the paper is organized as follows. First, the overall technologies for meet-
ing the management requirements of intelligent home network services are introduced
in Section 2. Then, our IHSM platform and details are presented in Section 3. Finally,
some concluding remarks and future works are drawn for the paper in Section 4.
Fig. 2. Open service structure enables home network service providers to accept both existing
services and new services at the same time through a common service platform [7]
Fig. 3. The intelligent home network service management (IHSM) platform consists of the
intelligent service management framework (ISMF) and the intelligent service management
system (ISMS), which were designed for an OSGi-based home gateway
We will continue to design and implement more details and make it practical. Our
major research areas in the future are on context aware-based service management
and home network security management.
References
1. National Computerization Agency: “The present and direction of digital home business”,
Digital home service model excavation workshop, (2003)
2. Kyu-Chang Kang, Jeon-Woo Lee: “Implementation of Management Agents for an OSGi-
based Residential Gateway”, Embedded S/W Technology Center, Computer and Software
Lab, ETRI(2003)
3. Choonhwa Lee, David Nordstedt, Sumi Held: “Enabling smart spaces with OSGi, Perva-
sive Computing IEEE, Volume-2 Issue-3, July-Sept, ZW3, pp.89-94 (2003) pp.89-94.
4. Open Service Gateway Initiative, About the OSGi Service Platform, Technical Whitepaper
Revision 4.1”, http: www.osgi.org (2005)
5. Richard S. Hall and Humberto Cervantes: “An OSGi Implementation and Experience Re-
port”, Laboratoire LSR Imag, rue de la Chimie Domain Universitaire, France(2004)
6. Haitao Zhang, Fei-Yue Wang, Yunfeng Ai: “An OSGi and Agent Based Control System
Architecture for Smart Home”, IEEE(2005)
7. Akihiro Tsutsui: “Management Architecture and Distribution Framework for Home Net-
work Services”, NGN Workshop(2005)
8. Daqing ZHANG, Xiao Hang WANG: “OSGi Based Service Infrastructure for Context
Aware Automotive Telematics”(2004)
9. Xie Li, Wenjun Zhang: “The Design and Implementation of Home Network System Using
OSGi Compliant Middleware”, IEEE Transactions on Consumer Electronics, Vol. 50,
No. 2, IEEE(2004)
10. Session Initiation Protocol(SIP), http://www.ietf.org/rfc/rfc2543, (1999)
11. Session Initiation Protocol(SIP), http://www.ietf.org/rfc/rfc3261, (2002)
COPS-Based Dynamic QoS Support for SIP Applications
in DSL Networks
Abstract. In this paper, several dynamic QoS solutions including Direct Diff-
Serv, Admission-based Direct DiffServ, Indirect DiffServ, and Hybrid DiffServ
are proposed to support largely emerging SIP-based P2P(Peer-to-Peer) and
ASP(Application Service Provider) multimedia applications in DSL networks,
most widely deployed as broadband access networks. The proposed solutions
are designed to be based on standard COPS protocol which is simple and ser-
vice-independent.
1 Introduction
SIP-based multimedia applications can be serviced in P2P(Peer to Peer) environment,
where each end user is responsible for the QoS provisioned, as well as in
ASP(Application Service Provider) environment, where the ASP is responsible for the
QoS. Dynamic QoS for a SIP multimedia application needs to be supported in an
appropriate way according to the corresponding service environment of the applica-
tion. Currently most QoS-enabled access networks are developed based on the Diff-
Serv IP QoS architecture because of the complexity problem of the other IntServ QoS
architecture[1,2,3]. In this paper, several dynamic QoS solutions including Direct
DiffServ, Admission-based Direct DiffServ, Indirect DiffServ, and Hybrid DiffServ,
are proposed to support largely emerging SIP-based multimedia applications in DSL
networks. How to apply each proposed dynamic QoS solution to the corresponding
service and network environment is also discussed in this paper.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 554 – 557, 2006.
© Springer-Verlag Berlin Heidelberg 2006
COPS-Based Dynamic QoS Support for SIP Applications in DSL Networks 555
offer and SDP answer exchange procedure embedded within SIP session establish-
ment[4,5], and perform DiffServ classification based on the collected QoS data. The
DiffServ classification is an environment-specific matter. (a) of Fig.1 shows an exam-
ple scenario of Direct DiffServ dynamic QoS solution for a simple SIP multimedia
application which does not include QoS preconditions.
(1) INVITE (1) INVITE (1) INVITE (1) INVITE (1) INVITE (1) INVITE
QoS data Collection and (2)
(2) 180 RINGING DiffServ Classification
(3) QoS data Collection and (3) 180 RINGING
DiffServ Classification (4) QoS data Collection and
(4) 180 RINGING DiffServ Classification
QoS data Collection and (5) 180 RINGING
(5)
(6) 180 RINGING DiffServ Classification (6) QoS data Collection and
(7) 200 OK (7) 180 RINGING DiffServ Classification
(8) COPS REQ (8) QoS data Collection and (9) 200 OK
DiffServ Classification
(10) COPS REQ
(9) COPS DEC
(10) 200 OK
(11) COPS DEC
(12) COPS REQ
(12) 200 OK
(13) COPS DEC (13) COPS REQ
(14) 200 OK
(14) COPS DEC
(15) ACK (15) 200 OK
Fig. 1. Example scenarios of Direct DiffServ, Admission-based Direct DiffServ, Indirect Diff-
Serv, and Hybrid DiffServ
Since Direct DiffServ does not require any additional signaling to support dynamic
QoS, it is very simple. But it does not provide any mechanisms to do admission con-
trol for IP QoS and policing service of the admitted IP QoS to protect authorized users
from non-authorized users. This means that it is difficult to apply Direct DiffServ
solution to the most broadband access environments where QoS policing is necessar-
ily required to filter non-authorized QoS packets. We believe that Direct DiffServ
solution will be useful to support P2P SIP multimedia applications, in which end
systems are fully responsible for QoS support, in enterprise network environments
where end users who are sending QoS packets can be trusted.
2) Admission-based Direct DiffServ solution adds some signaling mechanism be-
tween end-users and NSP to the Direct DiffServ solution. Through the signaling, NSP
can do admission control of IP QoS requests from end-users, and provide policing
service of the admitted IP QoS based on the authorization information given at the
signaling phase. (b) of Fig.1 shows an example scenario of Admission-based Direct
556 S. Park and Y. Choi
DiffServ solution to support dynamic QoS for simple SIP multimedia applications.
Since UA is performing the role of QoS PEP(Policy Enforcement Point) and
BRAS(Broadband Remote Access Server) of QoS-enabled DSL networks is acting as
QoS PDP(Policy Decision Point), in the Admission-based Direct DiffServ, the stan-
dard PEP-PDP COPS(Common Open Policy Service) protocol [6] can be used for the
QoS signaling between UA and BRAS. The PIB(Policy Information Base) defined as
a named object for the UA-BRAS COPS protocol will convey identification informa-
tion, QoS attribute values, and DiffServ classification information for each media
stream of a multimedia application. Though the UA-BRAS COPS signaling protocol
makes Admission-based Direct DiffServ solution more complex than Direct DiffServ
solution, it can provide policing service of the admitted QoS for authorized users to
protect from non-authorized users. Therefore, this solution will be very useful to sup-
port P2P SIP multimedia applications in most broadband access networks where QoS
policing is necessarily required.
3) In Indirect DiffServ solution, some QoS proxy servers of a multimedia application
will be responsible for supporting dynamic QoS on behalf of the end-user entities of
the application. SIP default(inbound/outbound) proxy servers will become QoS proxy
servers in QoS-enabled DSL access networks. COPS protocol can be also used for the
QoS signaling between default proxy server and BRAS, same as in the UA-BRAS
signaling of Admission-based Direct DiffSrev solution. (c) of Fig. 1 shows example
scenario of Indirect DiffServ solution to support dynamic QoS for simple SIP multi-
media applications. Default proxy servers, taking the role of QoS proxy servers, col-
lect identification information and QoS attribute values for each media stream and
determine DiffServ class, by capturing SDP offer and answer exchanged between
UAC and UAS via INVITE and 180 RINGING messages. When the remote default
proxy server receives 200 OK message indicating successful QoS negotiation from
UAS, it delivers the identification information and QoS attribute values to its remote
BRAS and requests IP QoS admission, by sending COPS REQ message. Remote
BRAS admits the IP QoS request by sending COPS DEC message, after checking the
configuration profile for the ASP and its resource allocation status. Admitted remote
default proxy server sends 200 OK response to local default proxy server. Local de-
fault proxy server, after receiving 200 OK message, performs similar IP QoS admis-
sion procedure with its local BRAS by using COPS protocol, and delivers the 200 OK
message to UAC, if successfully admitted. And then, BRAS will be ready to perform
packet classification and DSCP marking for media stream incoming from the user,
and aggregate queuing and prioritization.
In the Indirect DiffServ solution, BRAS is injection point of IP QoS and end-user
systems are not involved in the IP QoS enforcement. This means that Indirect Diff-
Serv can be easily deployed in the legacy QoS-unaware end-system environments. On
the other hand, Indirect DiffServ solution can be supported only in ASP environments
where there are QoS proxy servers and COPS signaling is supported between QoS
proxy server and BRAS. Moreover, Indirect DiffServ has a significant disadvantage
that there is no ways to support IP QoS in the access network ahead of BRAS because
end-systems are not involved in the QoS support. Consequently Indirect DiffServ
solution will be very useful to support dynamic QoS for ASP SIP multimedia applica-
tions in legacy broadband access networks where most end-systems are QoS-unaware.
COPS-Based Dynamic QoS Support for SIP Applications in DSL Networks 557
3 Concluding Remarks
In this paper, we proposed four different dynamic QoS solutions based on DiffServ.
Direct DiffServ solution is useful to support P2P SIP multimedia applications in en-
terprise network environments where end users can be trusted. Admission-based Di-
rect DiffServ can be easily applied to support P2P SIP multimedia applications in
most broadband access networks where QoS policing is necessarily required. Indirect
DiffServ solution will be effective in supporting dynamic QoS for ASP SIP multime-
dia applications in legacy broadband access networks where most end-systems are
QoS-unaware, since end-user systems are not involved in the IP QoS enforcement.
Hybrid DiffServ solves the problem of Indirect DiffServ solution that IP QoS is not
supported in access network ahead of BRAS by additionally allowing end-systems to
directly perform IP QoS enforcement activities such as DSCP marking.
References
1. Lakkakorpi, J., Strandberg, O., Salonen, J. : Adaptive Connection Admission Control for
Differentiated Services Access Networks. IEEE JSAC, Vol. 23, NO. 10 (Oct. 2005)
2. Salsano, S., Veltri, L.: QoS Control by Means of COPS to Support SIP-Based Applications.
IEEE Network (March/April 2002)
3. DSL-Forum TR-059 : DSL Evolution - Architecture Requirements for the Support of QoS-
Enabled IP Services (Sept. 2003)
4. Rosenberg , J. et al. : SIP : Session Initiation Protocol. IETF RFC 3261(June 2002)
5. Rosenberg, J., Schulzrinne, H. : An Offer/Answer Model with Session Description Proto-
col(SDP). IETF RFC 3264 (June 2002)
6. Durham, D. et al. : The COPS(Common Open Policy Service) Protocol. RFC 2748 (Jan.
2000)
IP Traceback Algorithm for DoS/DDoS Attack
1 Introduction
2 Related Works
There are three probabilistic packet marking methods such as node sampling,
edge sampling and advanced packet marking[3,4]. The node sampling method
requires many packets to reconstruct the attack path. The edge sampling method
(This research was supported by the MIC(Ministry of Information and Communi-
cation), Korea, under the ITRC(Information Technology Research Center) support
program supervised by the IITA(Institute of Information Technology Assessment)
(IITA-2005-(C1090-0502-0020))).
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 558–561, 2006.
c Springer-Verlag Berlin Heidelberg 2006
IP Traceback Algorithm for DoS/DDoS Attack 559
marks the result of the XOR operation with its own IP address and the IP address
of the previous router. This method is better than the node sampling method in
reconstructing the attack path. Advanced packet marking method provides an
authentication function during packet marking[6].
Multi-edge marking[5] appends the adjacent segment attack path to the record
route IP option of the packet as it travels through the network from attacker to
victim. Because of the limited space in the IP header, it can not mark all the
routers’ IP addresses into the record route IP option if the attack path is longer
than 9. If there are no other IP options in use, the record route IP option can
contain 9 IP addresses maximum.
When the edge router sets the marking flag to 1 with probability p, the record
route IP option is opened. The route data field is used for the XOR, EXT1 and
EXT2 fields. The XOR field stores the XOR value, the EXT1 and EXT2 fields
store some information to traceback. Each router has a Route Information Table.
This table consists of 2 field, packet ID and exclusive OR value. Packet ID field
is in use to store 16bits random value to distinguish the XOR value of packet.
Exclusive OR value field is in use to store XOR value that comes from the EXT2
field in the record route IP option of a packet.
Presently, the ratio of the fragmented packet of all Internet packets is less
than 0.25%[7]. Thus the identification field and the fragmentation field are not
560 H.-b. Yim and J.-i. Jung
used to identify which packet is fragmented. This paper uses the identification
field to store a 16bits random number to distinguish the XOR value of each
packet. To use this field, most 3bits of the fragmentation field are set to 010.
It means this packet is not fragmented. Most 3bits of the fragmentation field is
named fFlag.
In SPR algorithm, all routers that have received a pushback flag packet make
an information packet that includes the IP address of its own router. This infor-
mation packet is sent to the victim. This packet’s source IP address of IP header
is the router’s IP address and the destination IP address is the IP address of
the victim. This information packet sets the marking flag and pushback flag to
1. This packet is transmitted according to the normal routing algorithm. The
victim uses this packet to reconstruct the attack path. The initial TTL(Time To
Live) value of a normal packet is set to 255. The victim can know the number
of hops to the router using decrease of TTL value. The victim line up ascending
order of decrease of TTL value, is the order of the attack path from attacker to
victim. The ordering algorithm is simply the ascending order algorithm.
IP Traceback Algorithm for DoS/DDoS Attack 561
4 Simulation
In the simulation, we compare the following three traceback algorithms: the edge
sampling method, the multi-edge method and the proposed traceback method.
In the case of a one attacker attack the number of packets in the proposed
traceback scheme is decreased by an average of 16.5% compared to the multi-
edge method. In the case of a 10 attackers attack the number of packets in the
proposed traceback scheme is decreased by an average of 37.6% compared to
the multi-edge method. This result shows that the proposed scheme is a more
effective method than the multi-edge method in the case of DDoS attack.
Table 1. Required packet to reconstruct attack path when p=0.1
5 Conclusion
In this paper, we have proposed an effective packet marking algorithm to reduce
packets. This algorithm can be implemented in a router and uses record route
IP option and unused fields in the IP header. The marking algorithm uses a
probabilistic marking scheme and XOR operation to calculate IP addresses to
reduce the IP header packet size.
According to the simulation result, the proposed traceback algorithm is more
effective on a DDoS attack than the schemes available up to now. In the fu-
ture, we also need to test on a large scale network to check performance of this
algorithm in terms of the packet collection time.
References
1. Henry C.J. Lee, Vrizlynn L.L. Thing, Yi Xu, and Miao Ma, “ICMP Traceback with
Cumulative Path, an Efficient Solution for IP Traceback”, LNCS 2836, pp.124-135,
2003
2. “Internet Protocol : DARPA INTERNET PROGRAM PROTOCOL SPECIFICA-
TIO”, RFC791
3. K. Park and H. Lee. “On the effectiveness of probabilistic packet marking for IP
traceback under denial of service attack”, In Proc. IEEE INFOCOM 2001, page
338-347,2001
4. D. X. Song, A. Perrig, “Advanced and Authenticated Marking Scheme for IP Trace-
back”, Proc, infocom, vol.2, pp.878-886, 2001
5. Chen Kai, Hu Xiaoxin, Hao Ruibing, “DDoS Scouter : A Simple IP Traceback
Scheme” http://blrc.edu.cn/blrcweb/publication/kc1.pdf
6. Stefan Savage, David Wetherall, Anna Karlin and Tom Anderson, “Practical Net-
work Support for IP Traceback”, SIGCOMM, 2000
7. Ion Stoica, Hui Zhang, “Providing Guaranteed Services Without Per Flow Manage-
ment”, SIGCOMM, 1999
An Open Service Platform at Network Edge
Abstract. The last few years, users want various services that fit their needs and
preferences. Many services are provided at network edges and are increasing in
number. The open framework is needed for efficient service management and
uniform deployment of personal services at network edges. In this paper, we
propose the open service framework, which can be implemented as a platform
for the personal service at network edges by using Open Pluggable Edge Ser-
vice (OPES) concept. The proposed framework is composed of Databases, Pol-
icy Repository, Rule Manager, Policy Adapter, Admission Manager, Operations
Support System (OSS) Adapter and OPES-based Service Delivery Manager for
the deployment of personalized and QoS guaranteed services in a standard way.
In order to perform the feasibility of this platform, we have implemented a sim-
ple example and shown some its results.
1 Introduction
Personalized service is to adapt the provisioned services to fit the needs and prefer-
ences of a user or a group of users in a static or dynamic way. The process of person-
alized service gathers information of user, device and service through their interac-
tions and in turns, stores the collected raw information into database.
The edge service means service for deploying service by 3rd-party Service Devel-
opers and Network Provider at network edge. Because each edge has many services, it
needs the method to manage and deploy services. In [1], a platform is proposed for
managing and deploying various services, but it does not mention on the concrete
relationship with personalized services.
So we propose the open service framework, which can be represented as the per-
sonalized service platform with the network and service control functionality for the
dynamic provision of a variety of edge-based services, being deployed and managed.
Section 2 introduces the Open Pluggable Edge Service (OPES) architecture to be
applied in the framework. Section 3 describes the proposed framework for deploying
the personalized services in an effective way and section 4 implements a case of per-
sonalization service for the feasibility of the proposed framework. Section 5 provides
conclusive remarks and directions for future work.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 562 – 565, 2006.
© Springer-Verlag Berlin Heidelberg 2006
An Open Service Platform at Network Edge 563
services for HTTP. The framework is a one-party consent model in which each ser-
vice is authorized explicitly by at least one of the application-layer endpoints.
The OPES, existing in the network edge, provides services that modify requests,
modify responses and create responses. The architecture of OPES can be described by
OPES entities, OPES flows and OPES rules. An OPES entity residing inside OPES
processors is an application that operates on a data flow between a data provider ap-
plication and a data consumer application.
An OPES entity consists of an OPES service application and a data dispatcher. An
OPES service application can analyze and transform messages on data stream. A data
dispatcher invokes an OPES service application according to an OPES rule-set and
application-specific knowledge. If there is no needed service application in local
OPES processor or is useful for OPES processor to distribute the responsibility of
service execution in an OPES service application of remote Callout Servers, the data
dispatcher invokes remote service application by communicating one or more Callout
server(s). In this case, OPES Callout Protocol (OCP) is used for communication be-
tween a data dispatcher and a Callout Server.
OPES flows are data flows among a data provider, a data consumer and one or
more OPES processor(s) in which a data dispatcher must be existed. The exchanges
of data between a data provider and a data consumer are independent of protocols. For
the simplicity and implementation of proposed platform, we select HTTP, a basic
protocol of Web Services, as the example for the underlying protocol in OPES flows.
The rule-set is a superset of all OPES rules. The OPES rules consist of a set of
conditions and their related actions which can be specified as when and how to exe-
cute OPES services on data stream. The data dispatcher examines the rules and
invokes service application for offering the service at the points identified by the
numbers 1 through 4.
policies with their priorities to the RM. The AM can perform the functionality of
whether the requested application service can satisfy the negotiated QoS require-
ments. In order to perform this functionality, the AM can access the OSSA, which
offers the network and service management information for representing the status of
network resource and service provision. Therefore, the OSDM can deploy many suit-
able application services to which the received rules from the RM are applied. Also, it
enforces the established policies related to application services into QoS-aware net-
work nodes such as edge router, access node and access switch.
5 Conclusion
We propose the open service framework, which can be represented as the personal-
ized service platform with the network and service control functionality for the dy-
namic provision of a variety of edge-based services, being deployed and managed. In
order to make this framework, we design the system components such as Databases,
Policy Repository, Rule Manager, Policy Adapter, Admission Manager, OSS Adapter
and OPES-based Service Delivery Manager.
However there are many Service Providers at the network edge and they want to
provide service at other Network Provider edge. Network Provider requires Service
Providers to comply with its API for deploying their services. For following the con-
venience and standard to Service Provider, we will extend the proposed framework
towards Service Delivery Platform (SDP).
References
1. Falchuk, B., Chiang, J., Hafid, A., Cheng, Y.-H., Natarajan, N., Lin, F.J., Cheng, H.: An
open service platform for deploying and managing services at network edges, Open Archi-
tectures and Network Programming, pp. 77 – 86, April 2003.
2. OPES Working Group: http://www.ietf.org/html.charters/opes-charter.html.
3. Dinesh C. Verma: Policy-Based Networking-Architecture and Algorithms, New Riders,
2000.
4. Timo Laakko, Tapio Hiltunen: Adapting Web Content to Mobile User Agents, IEEE Inter-
net Computing, Vol 9. No 2, pp. 46-85, 2005.
5. Panayiotou, C., Andreou, M., Samaras, G., Pitsillides, A.: Time based personalization for
the moving user, Mobile Business, pp. 128 – 136, July 2005.
6. Dong-Jun Lan, Chun Ying, Jun Liu, Wei Lu : Policy Lifecycle and Policy Enabled Telecom
Open Service Delivery Platform, IEEE International Conference on Web Service, 2005.
7. Sun Microsystems.Inc: http://java.sun.com/products/sjwtoolkit/index.html.
8. Deitel: Java Web Service for Experienced Programmers, Prentice Hall , 2002.
Hybrid Inference Architecture and Model
for Self-healing System*
1 Introduction
A computer system would satisfy the requirements of autonomic computing, if the
system can configure and reconfigure itself by knowing the operating environments,
protect and heal itself from various failures or malfunctions. The core of autonomic
computing, a recently proposed initiative towards next-generation IT-systems capable
of ‘self-healing’, is the ability to analyze a data in real-time and to predict potential
problems [1]. Currently, most self-healing systems perform healing after error
occurrence [2]. Sun has developed a new architecture for building and deploying
systems and services capable of Predictive Self-Healing [3]. However, this is
predicted through limited elements, because of healing in the same manner as IBM
(vender-dependant). In order to know the environments and detect failure, an
autonomic system needs the capability of acquiring the information through self-
monitoring.
*
This work was supported in parts by Ubiquitous Autonomic Computing and Network Project,
21th Century Frontier R&D Program, MIC, Korea, ITRC IITA-2005-(C1090-0501-0019),
Grant No. R01-2006-000-10954-0, Basic Research Program of the Science & Engineering
Foundation, and the Post-BK21 Project.
**
Corresponding author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 566 – 569, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Hybrid Inference Architecture and Model for Self-healing System 567
In this paper, a hybrid inference model is proposed to solve this problem. These
proactive prediction and probing capabilities will provide the system management
components with the pertinent information such that self-healing are possible for
critical system resources. These models adopt a selective model, depending on the
system situation, for self-diagnosing and prediction of problems. In this paper
characteristics of each algorithm are detailed through a comparison of four inference
model and a demonstration by experiment.
The remainder of this study is organized as follows: Section 2 describes the
proposed architecture and model; Section 3 describes an evaluation through its
implementation and experiments; and Section 4 presents the conclusion.
Triple Modular Redundancy (TMR) [7] is method used for increasing fault tolerance
in safety critical software system. When four prediction algorithms can’t select a
suitable algorithm in specified situation, the Monitoring Agent predicts the system
situation by the comparison among each algorithm using TMR. TMR is built up from
three equal components (all three components should produce the same output for the
same input) and an output selector. Four algorithms is flexibly changed and replaced
according to the system situation. TMR assumes that the algorithms must have equal
input and output. If the one algorithm has different output, it is recognized as a fault.
And the others are recognized as a correct value of inference.
Fig. 1. Error rate changed by the time flow Fig. 2. A processing time to predict the sys-
tem situation by the number of error
In the first experiment, comparison of the existing system was performed, that only
used the ID3 to predict the proposed system. As a result, it was concluded that a rate
of error log is reduced through use of the suitable inference model according to
system requirements. The above experiment result is presented in Figure 1.
Hybrid Inference Architecture and Model for Self-healing System 569
The time taken for prediction in the second experiment was measured in Figure 2.
First, we assume a learned data has an enough quantity. It proposed model’s
efficiency was proven through comparison of the number of errors occurring in the
system. If the number of error is fewer, ID3's result displays similar performance such
as the result of applying a hybrid model. However, the number of errors was high, and
the prediction time of the hybrid model, which predicts the characteristic of the
system as well as resource information, was shorter.
4 Conclusion
Previous monitoring technologies of the self-healing system dissatisfied a requirement
of the ubiquitous computing. In recent years, numerous studies have attempted to find
and explore self-healing systems, such as IBM [2], and Sun [3]. However, it is a
healing of vender-dependent and the use of a single inference model. This paper
proposed a hybrid inference model to recognize operating environments and predict
error occurrence. Therefore, we can perform an efficient prediction by the system’s
status and environment through a hybridization of inference model. Four algorithms
proposed in this paper are clearly divided with characteristics and can be used in
hybrid inference models, according to the system request. We designed architecture to
support hybrid inference model for self-healing system. We compare an efficiency of
algorithm by experiment about the time of prediction, the correctness of prediction
and system’s load. So, we made up for the weak point in our previous work [4].
References
1. R.K. Sahoo, A. J. Oliner, I.Rish, M. Gupta, J.E. Moreira, S. Ma, “Critical Event Prediction
for Proactive Management in Large-scale Computer Clusters”, ninth ACM SIGKKD
international conference on Knowledge discovery and data mining, pp. 426-435, 2003
2. B. Topol, D. Ogle, D. Pierson, J. Thoensen, J. Sweitzer, M. Chow, M. A. Hoffmann, P.
Durham, R. Telford, S. Sheth, T. Studwell, “Automating problem determination: A first
step toward self-healing computing system”, IBM white paper, Oct. 2003
3. Sun Microsystems: Predictive Self-Healing in the Solaris 10 Operating System,
http://www.sun.com/bigadmin/content/selfheal
4. Jeongmin Park, Giljong Yoo and Eunseok Lee, "Proactive Self-Healing System based on
Multi-Agent Technologies", ACIS International Conference on Software Engineering
Research, Management & Application(SERA 2005), IEEE, pp.256-263, Aug.2005
5. Kwang H.Lee, ‘First Course on Fuzzy Theory and Applications’, Advances in Soft
Computing, Springer, 2005
6. Sucheta Nadkarni, Prakash P. Shenoy, “A causal mapping approach to constructing
Bayesian networks”, Decision Support Systems, Vol.38, pp.259-281, Nov.2004
7. J. Von Neumann, ‘Probabilistic logics and synthesis of reliable organisms from unreliable
components’ in Automata Studies, C. E. Shannon and J. McCarthy, Eds. Princeton, NJ:
Princeton Univ. Press, pp. 43–98, 1956
A Node Management Tool for Dynamic Reconfiguration
of Application Modules in Sensor Networks*
1 Introduction
Researchers have done much work on wireless networks for a number of years and
have developed fairly sophisticated sensor network systems in the face of strict con-
straints such as low power, low cost, small size, fault tolerance, flexibility, and secu-
rity. The reconfiguration of application modules for sensor networks is challenging
about above constraints at each node. Reconfiguration and self-adaptation are impor-
tant factors of sensor networks that are required to operate in dynamic reconfiguration
that includes functional changes and nonfunctional performance improvements.
Dynamically adaptive application comprises tasks that detect internal and external
changes to the system, reflecting the new application environment conditions. Wire-
less sensor networks, in particular, require reconfigurable capabilities that enable
them to handle a multitude of nodes. Consequently, existing systems that collect and
aggregate data in sensor networks provide an active tool for managing each node.
Our approach applies the domain of robust, fault-tolerant embedded systems. It is
able to reconfigure various changes tailored for application of nodes. The system can
quickly find a new configuration and adapt to environment changes. This paper pro-
posed a systematic node management tool of node application modules using the
minimum memory resources in the sensor nodes. Based on this approach, the tool
uses bug modifications, function updates, and application changes on the nodes. This
tool is implemented based on Nano-Qplus[6], developed by ETRI, Korea. If we util-
ize the proposed tool, therefore, developers develop application modules of the nodes
simultaneously and we can flexibly deal with changes on the node.
*
This work was supported by the Soongsil University Research Fund.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 570 – 573, 2006.
© Springer-Verlag Berlin Heidelberg 2006
A Node Management Tool for Dynamic Reconfiguration 571
2 Related Works
The sensor network is defined that it can collect the necessary sensing information
and each node has a processor that can process collected information and wireless
telecommunication device that can transmit it. Typical sensor networks consist of
tens, if not hundreds, of nodes, including special sink nodes that connect them to
global networks, such as the Internet. Communication occurs regularly over multiple
hops, and due to frequent poor link quality, reliable data collection at the sink node is
a significant problem. Namely, the sensor networks will provide the bridge between
the sensors and the physical world due to their ability to observing and controlling the
sensing data in real-time. The above described features ensure a wide range of appli-
cations for sensor networks[1]. There were some researches in relation to sensor net-
work operating systems such as TinyOS[2], SOS[3], Maté [4], MANTIS[5] and
Nano-Qplus[6]. Especially, Nano-Qplus, our approach target OS, supports ultra-small
kernel size(smaller than 10KB), distributed, real-time, and smart operating system
that is suitable for various application areas such as medical care, environment, disas-
ter prevention, digital home, national defense, and industrial equipment.
Prevalent sensor network operating systems have some problems as follows: (1)
use only version management system tools by targeting simultaneous development,
(2) do not support automatic version synchronization between development environ-
ment nodes and real nodes, (3) rarely provide the fine-grained modular update to fit
the given memory limit or to optimize the system.
3 System Architecture
We have prototyped an architecture for software reconfiguration shown in Figure 1. It
was a better way to update a sensor network using a centralized database. It is possi-
ble to increase the reusability, self-adaptability, and flexibility. However, we proto-
typed our approach with the current limitation of the Nano-Qplus such as high update
cost caused by reconfiguring a full application image and testing in a smaller scale
sensor networks. We have plans to extend the modular update referred to SOS to-
wards the modular-reconfigurable Nano-Qplus environment and to verify effective-
ness and constraints such as energy consumption and memory limitation.
The NVSync(Nano-Qplus Version Synchronization), a node management tool
primarily requires loading of a prerequisite image on an initial target node through the
serial line and then gathers the node version information on the node and that on the
repository. It updates the version information stored in the node information map for
each node and reconfigures the application image using the target node information,
when the role of the target node is changed. Most of the energies are wasted when it
gathers the sensing data in the sensor field. The application module of a sensor node
can be modified to gather proper data using the proposed approach with energy effi-
ciency, when the sensor field is changed. Moreover it is possible to modify the appli-
cation modules of a node without removing the existing node, when the application
572 S. Jung et al.
Fig. 1. System architecture(left fig.) and system behavioral sequence (right fig.) for the
NVSync
Component Descriptions
NVSync UI ƒ Displays entire user interface of SNM.
Version Synchro ƒ Registers/changes new or modified image’s node information in the node information map from
nization Manager the Version Manager.
ƒ Maps/remaps the node information map and the sensor network nodes.
ƒ Executes the version reconfiguration after comparing the node version information of the Versio
n Manager and the node information of the Node Monitor on the node.
ƒ Applies the predefined rules of each node from the Rule Manager to reconfiguration.
Rule manager ƒ Selects the desired reconfiguration node.
ƒ Sets up the predefined rule consists of 3 types – user mode, periodic mode, and direct mode.
Version Synchro
ƒ A centralized database for the node information map.
nization Engine
ƒ Primarily load initial application image built from Image Builder in the development environmen
Program Loader
t.
Remote Loader ƒ Loads a modified application image through the wireless networks.
ƒ Stores version information of each node and periodically sends it to the Version Synchronization
Node Monitor
Manager of the NVSync.
ƒ consists reconfigurable environment of application images and ports updated images through the
Bootloader
Wireless Communication Components.
Version Synchro ƒ synchronizes updated image and records the event logs showing whether update is update compl
nization Agent ete or incomplete.
References
1. Ian F. Akyildiz, Weilian Su, Yogosh Sankarasubramaniam, and Erdal Cayirci, “A Survey on
Sensor Networks,” IEEE Communications Magazine, pp.102~114, Aug. 2002
2. Hill. J., Szewczyk. R., Woo. A., Hollar. S., Culler. D., and Pister. K. “System architecture
directions for networked sensors,” In Proceedings of the ninth international conference on
Architectural Support for Programming Languages and Operating Systems (ASPLOS 2000),
ACM Press, pp.93-104
3. Chih-Chieh Han, Ram Kumar, Roy Shea, Eddie Kohler and Mani Srivastava, “A Dynamic
Operating System for Sensor Nodes,” In proceedings of 3rd International Conference on
Mobile Systems, Applications, and Services(MobiSys 2005) on the USENIX, pp.163-176,
2005.6
4. Levis. P., and Culler. D., “Maté: A tiny virtual machine for sensor networks,” In Interna-
tional Conference on Architectural Support for Programming Languages and Operating Sys-
tems, San Jose, CA, USA (Oct. 2002).
5. Abrach. H., Bhatti. S., Carlson. J., Dai. H., Rose. J., Sheth. A., Shucker. B., Deng. J., and
Han. R., “MANTIS: system support for multimodal networks of in-situ sensors,” In Pro-
ceedings of the 2nd ACM international conference on Wireless sensor networks and appli-
cations (2003), ACM Press, pp. 50-59
6. Introduction and download of Nano-Qplus, http://www.qplus.or.kr
Path Hopping Based on Reverse AODV for Security
1 Introduction
A mobile ad hoc network is a dynamically self-organizing network without any cen-
tral administrator or infrastructure support. If two nodes are not within the transmis-
sion range of each other, other nodes are needed to serve as intermediate routers for
the communication between the two nodes [1, 2].
In ad hoc wireless networks, transmitted data is susceptible to potential attacks.
Eavesdroppers can access secret information, violating network confidentiality.
Hackers can directly attack the network to drop data packets, inject erroneous mes-
sages, or impersonate as a member node. To increase security, physical protection of
the network from malicious node is important.
In this study we propose path hopping based on reverse AODV [2]. In R-AODV,
which is an easy multipath searching method, destination node uses reverse RREQ to
find source node rather than a unicast reply. It reduces path fail correction messages
and also source node builds partial or complete non-disjoint multipath from source to
destination. Hopping paths means source node sends each data packet through differ-
ent paths each time, therefore eavesdropper will not get whole data and also its intru-
sion to network become harder [3-7].
Physical protection of data from malicious invader is an important security method.
It can decrease or prevent packet loss by active malicious nodes [7].
*
Correspondence author.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 574 – 577, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Path Hopping Based on Reverse AODV for Security 575
ρ m = (N p • M ) / N all . (1)
576 E. Talipov et al.
0.5
0.4
0.3
0.2
0.1
0
1 2 3 4 5 6 7 8 9 10
Number of Paths
3 Performance Results
We describe the simulation environment used in our study and then discuss the results
in detail. Our simulations are implemented in Network Simulator (NS-2) [6].
We compare performance of AODV, R-AODV and PHR-AODV. Figure 4 shows
packet deliver ratio of each protocol. R-AODV has better delivery ratio than other
protocols have. PHR-AODV delivery ratio is less than other protocols, because it
maintains more paths than others. Figure 5 shows the control packet overhead for
setting routing paths. PHR-AODV has less packet overhead than that of R-AODV.
100.00 25000
90.00
20000
80.00
Deliver Ratio (%)
Control Packets
15000
70.00
60.00 10000
50.00
5000
40.00
30.00 0
10 20 30 40 50 75 10 20 30 40 50 75
Number of nodes
Number of nodes
Fig. 4. Packet Delivery Ratio, when the num- Fig. 5. Control Packet Overhead, when num-
ber of nodes varies ber of nodes varies
Path Hopping Based on Reverse AODV for Security 577
Figure 6 shows the average remained energy of each protocol. Figure 7 shows en-
ergy difference to express the distribution rate. PHR-AODV has less energy differ-
ence and balanced energy than others.
90.00
12.00
88.00
86.00
Energy (out of 100)
10.00
84.00
Difference
8.00
82.00
80.00 6.00
78.00
4.00
76.00
74.00 2.00
10 20 30 40 50 75 10 20 30 40 50 75
Number of nodes Number of node
Fig. 6. Average energy remained, when num- Fig. 7. Energy Difference, when number of
ber of nodes varies nodes varies
4 Conclusions
Security is a significant issue in ad hoc networks. Intrusion of malicious nodes may
cause serious impairment to the security. To decrease effect of malicious nodes, we
proposed the idea of path hopping based on reverse AODV, in which the source node
attempts to hop among available paths and split data. We conducted extensive ana-
lytic model and a simulation study to evaluate the performance of PHR-AODV with
the R-AODV and AODV using NS-2. The results show that PHR-AODV maintains
reasonable packet delivery ratio, energy consumption and energy distribution while
increasing security of network. Our future work will focus on studying practical de-
sign and implementation for PHR-AODV.
References
1. C. Perkins, E. Belding-Royer Ad hoc on-Demand Distance Vector (AODV) Routing, RFC
3561, July 2003
2. Chonggun Kim, Elmurod Talipov, and Byoungchul Ahn, “A Reverse AODV Routing Pro-
tocol in Ad Hoc Mobile Networks” , LNCS 4097, pp. 522 – 531, 2006.
3. C. K.-L. Lee, X.-H. Lin, and Y.-K. Kwok, “A Multipath Ad Hoc Routing Approach to
Combat Wireless Link Insecurity,” Proc. ICC 2003, vol. 1, pp. 448–452, May 2003.
4. S.-J. Lee and M. Gerla, “Split Multipath Routing with Maximally Disjoint Paths in Ad Hoc
Networks,” Proc. ICC 2001, vol. 10, pp. 3201–3205, June 2001.
5. M. K. Marina and S. R. Das “On-Demand Multi Path Distance Vector Routing in Ad Hoc
Networks,” Proc. ICNP 2001, pp. 14– 23, Nov. 2001.
6. NS, The UCB/LBNL/VINT Network Simulator (NS), http://www.isi.edu/nsnam/ns/, 2004.
7. Zhi Li and Yu-Kwong Kwok, “A New Multipath Routing Approach to Enhancing TCP
Security in Ad Hoc Wireless Networks” in Proc. ICPPW 2005.
Mixing Heterogeneous Address Spaces in a
Single Edge Network
Abstract. The growth of IPv4 Internet has been facing the infamous
IP address depletion barrier. In practice, typical IPv4 Internet edge net-
works can be expanded by incorporating private addresses and NAT
devices.
In this paper, major limitations of NAT-expanded private networks
are presented. Furthermore, a solution is proposed to encourage the
mixed usage of private and public IP addresses in a single edge network
domain. The solution comprises of two key ideas : super-subnet mask and
shared NAT. Super-subnet mask removes the routing boundary between
private and public hosts. Shared NAT saves public IP address resources
by sharing them among several private networks. These ideas not only
encourage the coexistence of heterogeneous address classes, but also lead
to efficient sharing of global IP addresses.
1 Introduction
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 578–581, 2006.
c Springer-Verlag Berlin Heidelberg 2006
Mixing Heterogeneous Address Spaces in a Single Edge Network 579
Mixing private and public hosts can incur cumbersome problems with regard to
arrangement and inter-operability.
1. The NAT hinders private and public hosts from reaching to each other, by
separating a premises network into two different routing domains.
2. The NAT consumes at least one public address per one premises network
and the occupation tends to be long.
3. The NAT limits the performance of not only inter-premises but also intra-
premises network. The public host must content with private hosts for the
NAT(and routing) resources, in order to reach an intra-premises host.
The most straightforward solution to the first problem – the lack of direct routes
– is to assign an oversized subnet mask to local hosts. If a network address is
large enough to cover all the private and public addresses in the premises, the
subnet mask for the oversized network address is called a super-subnet mask1 .
Edge Router
211.100.1.2
ISP Edge Network 211.100.1.1 Netmask = 255.255.0.0
211.100.1.0/24
Aggregation Switch Shared NAT
192.168.1.1
...... Netmask = 255.255.255.0
Subscriber Lines
Switch/Hub Switch/Hub
...... ......
192.168.1.11 192.168.1.12 211.100.1.31 211.100.1.32
When a private host wants to communicate with the global Internet hosts, its
address should be translated into a public one by a NAT. Because the super-
subnet organization removes the logical boundary between each premises net-
work, a single NAT server is sufficient to serve all the private hosts within the
edge network.
A shared NAT serves the private hosts with a route to the global Internet, in
the similar way as an edge router serving the local public hosts. A large number
of individual NATs can be replaced by only one shared NAT. Hence, the total
number of public IP addresses occupied by NATs can be reduced significantly
across an edge network.
Destination
Source Local Public Local Private Foreign
Local Public N/A keep silent edge router
Local Private keep silent N/A shared NAT
Foreign bypass bypass N/A
3 Conclusion
The advantage of super-subnet can be summed up as follows.
Compared to the legacy NAT-expanded networks, public IP addresses are less
wasted and need not to be occupied all the time. The public IP resource can be
saved for more precious servers instead of wasted by individual NATs.
Mixing Heterogeneous Address Spaces in a Single Edge Network 581
All hosts within the super-subnet can directly communicate via L2 links. The
inefficiency caused by individual NATs is ameliorate by direct routing between
local hosts.
High performance bridge-based residential gateways can be implemented with
lower costs compared to NAT-based ones. Contrary to that NAT-based RGs
should be loaded with powerful processor and complicated softwares, bridge-
based RGs can be implemented with a few simple extensions to the L2 switches.
They are more competitive to the simple L2 switches than NAT-based ones, with
respect to the performance.
4 Related Works
There are some related works with regards to address reuse and NAT extensions.
MobileNAT [3] utilizes the NAT techniques to facilitate the mobility of wire-
less IP networks. The extensive use of DHCP is also applicable to other NAT
related works.
As a method for address sharing that exhibits more transparency than NAT,
RSIP [4] is proposed and published as RFC 3103 [5]. RSIP requires hosts to be
modified in order to interact with RSIP gateways.
Another point of view on the NAT problems is that it makes the peers con-
fused about their identities by implicitly transforming the address headers. As
a consequent, new routing and tunneling protocols that play the similar role as
NAT are proposed. A few explicit identity-based routing mechanisms are known
so far [6] [7] [8].
References
1. Srisuresh, P., Egevang, K.: Traditional IP Network Address Translator (Traditional
NAT). RFC 3022 (Informational) (2001)
2. Srisuresh, P., Holdrege, M.: IP Network Address Translator (NAT) Terminology
and Considerations. RFC 2663 (Proposed Standard) (1999)
3. Buddhikot, M., Hari, A., Singh, K., Miller, S.: Mobilenat: a new technique for
mobility across heterogeneous address spaces. In: WMASH ’03: Proceedings of the
1st ACM international workshop on Wireless mobile applications and services on
WLAN hotspots, New York, NY, USA, ACM Press (2003) 75–84
4. Borella, M., Montenegro, G.: Rsip: Address sharing with end-to-end security (2000)
5. Borella, M., Grabelsky, D., Lo, J., Taniguchi, K.: Realm Specific IP: Protocol Spec-
ification. Internet Engineering Task Force: RFC 3103 (2001)
6. Ramakrishna, P.F.: Ipnl: A nat-extended internet architecture. In: SIGCOMM ’01:
Proceedings of the 2001 conference on Applications, technologies, architectures, and
protocols for computer communications, New York, NY, USA, ACM Press (2001)
69–80
7. Turányi, Z., Valkó, A., Campbell, A.T.: 4+4: an architecture for evolving the inter-
net address space back toward transparency. SIGCOMM Comput. Commun. Rev.
33(5) (2003) 43–54
8. Walfish, M., Stribling, J., Krohn, M., Balakrishnan, H., Morris, R., Shenker, S.:
Middleboxes no longer considered harmful. MIT Technical Report TR/954 (2004)
Delivery and Storage Architecture for Sensed
Information Using SNMP*
Abstract. Many researches on context aware computing are carried out around
the world. Among them, Context-Toolkit and Semantic Space provide separation
of concerns between sensor and application. They make application developing
easier. However, they have one problem that is lacking of simplicity in
communication, compatibility and flexibility in building systems. To solve it, we
propose one delivery and storage structure using standardized simple network
management protocol which is useful to deliver, store and manage sensed
information. We also verify that this architecture is efficient in wireless sensor
network to deliver and store environmental information through an
implementation of a SNMP agent. We confirm that this architecture with
simplicity, compatibility and flexibility gives the efficiency to developing
systems.
Keywords: sensed information management, sensor network management,
SNMP.
1 Introduction
Many researches about context aware computing are carried out around the world.
Among them, Context-Toolkit[1] and Semantic Space[2] use a medium to manage
context information between the sensor and application in order to resolve the
problem of dependency. However, to let the application use context information, the
medium in between must be newly established, or a complex connection must be
made with the previous medium.
In Context-Toolkit, there is a problem of increasing connection points for
communications. An example for this can be the understanding of overall resources -
Toolkit must execute resource discovery to understand the overall resources.
The widget of Context-Toolkit has a communication module and data-processing
module, which are vulnerable to the application and communication protocol of
*
This research was supported by the Program for the Training of Graduate Students in
Regional Innovation which was conducted by the Ministry of Commerce Industry and
Energy of the Korean Government.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 582 – 585, 2006.
© Springer-Verlag Berlin Heidelberg 2006
Delivery and Storage Architecture for Sensed Information Using SNMP 583
SNMP agent of host PC, which collects and manages environment sensed
information. And SNMP manager of outside systems, which can approach to
information, approaches to SNMP agent.
SNMP agent module is consisted of SNMP agent, data transmittance, and MIB.
SNMP agent manages and controls sensor. MIB stores environment information of
sensor. SNMP manager module is consisted of SNMP manager and MIB.
In this paper, we compose server side's system (agent) to confirm that SNMP agent
could serve environment information from wireless environments to client side's
system (manager). Below Fig. 2 shows composition of agent part.
The sensor modules sense from environments and deliver sensed information to
other sensor modules or the base one through wireless communications. Sensor
modules install application programs managed by TinyOS. In this paper, we measure
temperature, illumination, internal voltage etc. by installing a OscilloscopeRF
584 D. Choi et al.
application program. Except the base sensor module, the others are installed in
OscilloscopeRF program that could send the sensed values. The base sensor module is
installed in base station program.
The base sensor module acting like a gateway passes the sensed information
acquired by the UART using Java interface to the SNMP agent application module.
We composed the system that passes the sensed information using general MoteIF
application program here.
Environment information accepted in this host PC was passed to a SNMP agent
part and the sensed values was managed by the SNMP agent. After monitored by an
automatic polling program, we received the result managed by SNMP agent in this
sensor information table, Fig.3. The periodic polling test shows that the SNMP agent
brings environment information from the sensor properly.
The SNMP manager acts like a client collects environment information from the
SNMP agent and manages sensors through these commands Trap, Get, Set etc.
SNMP is widely used all round as a standard protocol, and the several application
programs are possible. SNMP manager can access to the SNMP agent through basic
information (IP, community information, MIB OID etc.) that can approach to the MIB.
Simply inserting SNMP manager module to an application or a middleware will
get the environmental information of the sensor easily and the information will be
able to use. Fig.4 is the environment information of sensor got from MIB browser
(client).
Delivery and Storage Architecture for Sensed Information Using SNMP 585
Through the experiments above, I confirm that the SNMP module could search and
control the environment information of sensor.
The wrapping of the agent using SNMP provides a transparent characteristic to the
developers of application programs who do not need to know every vendor's API in
input processing of various sensors. This gives the efficiency to developing systems.
3 Conclusion
In order to manage sensed information that supports wireless sensor network, this
paper suggests a method of transmitting and saving sensed information through
SNMP protocol. Also, through realizing agent parts, it suggests that the structure
using previous SNMP protocol is also efficient in transmitting and saving sensed
information. Through using such architectures, a sensor management structure can be
built to satisfy the functions of information transmission, information saving,
conditional informing and control.
By using a standardized SNMP protocol, problems in previous context aware
computing technologies can be resolved, such as resource discovery, incompatibility
and singleness of communication type, insufficient flexibility in development, etc.
References
1. Anind K. Dey, "Providing Architectural Support for Building Context-Aware Applications,"
PhD thesis, College of Computing, Georgia Institute of Technology, Dec 2000.
2. X. Wang, D. Zhang, J. S. Dong, C. Chin and S. R. Hettiarachchi. "Semantic Space: A
Semantic Web Infrastructure for Smart Spaces", IEEE Pervasive Computing, 3(3):32-39,
July-September 2004.
3. William Stallings, "SNMP, SNMPv2, SNMPv3, and RMON1 and 2 Third Edition,"
Addison Wesley New York.
GISness System for Fast TSP Solving and
Supporting Decision Making
Abstract. This paper shows that geographic information system can success-
fully solve TSP (travelling salesman problem). It has been done using a module
of the designed and implemented by authors GISness system. Three algorithms
for solving TSP are available, including the proposed by authors an hybrid
algorithm called 2optGAM. The results of research show that the algorithm is
very promising.
1 Introduction
Nowadays business organizations are gathering data about sales, customers and
demographic profiles. Most of the collected data are geographic data with two
important features: a value and its location in space. It was estimated in [5] that more
than 70% of trade data is spatially oriented. Geographic information systems (GIS)
allow collecting, manipulating and analyzing data in a spatial context.
The designed and implemented by authors GISness (GIS for Business) system is
GIS system with elements of DSS (Decision Support System). A set of possible
decisions is created as a result of user interaction with system. GISness offers three
different views on geographic information: (i) the database view – every object on a
digital map has a corresponding entry in DB, (ii) the map view – data are presented in
form of interactive maps; maps allow querying, editing and analysing information,
(iii) the model view – a sequence of functions used to transform data into information
together with sets of data form a cartographic model [4, 5].
Since the TSP is spatially oriented, geographic information systems can serve as a
sophisticated environment for TSP data manipulation and visualization [1, 10]. We
created a module of GISness for solving TSP. Three approaches were adapted,
including the well-known nearest neighbour algorithm, 2opt method [8], and our own
algorithm 2optGAM being an improvement of the 2optGA algorithm [9].
2 GISness System
GISness has been developed in document-view architecture, where the document is a
digital map containing information about objects. This map contains spatial data and
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 586 – 589, 2006.
© Springer-Verlag Berlin Heidelberg 2006
GISness System for Fast TSP Solving and Supporting Decision Making 587
descriptive data. One of the main features of GISness is the possibility of linking
database to the map. This distinguishes GISness from GIS-systems which use static
maps and other graphic applications, e.g. CAD. GISness system stores descriptive
data in database governed by its own efficient management system GISnessDB.
Objects can be represented by attributes defined by the user as a quantity e.g. a
number of inhabitants, or as a text e.g. a name of a town. GISness allows exchanging
descriptive data with other database systems through import and export data.
The system was designed with UML standard supported by Power Designer 9.0 in
Sybase environment [6]. It has been developing since 2004 as light but robust system
[4]. Recently, it is composed of 21 000 lines of C++ code that create 82 classes.
GISness provides user with functions for descriptive data analysis: classification
and/or selection of objects depending on their features, performance of calculations
on attributes, viewing on attributes statistics and graphs, and for spatial data analysis:
calculation of distances between objects, calculation of statistics depending on
location and objects relations, and visualization of results of analysis.
The nearest neighbour [8] algorithm starts in a chosen city and in turn selects the
closest unvisited city until all cities are in the route. Solutions often contain crossing
edges and the length of route depends on a chosen starting point. GISness computes
routes for each city taken as starting point and returns the shortest one as the solution.
The 2opt method returns local minima in polynomial time [8]. It improves route
by reconnecting and reversing order of subroutes. Every pair of edges is checked
588 I. Pozniak-Koszalka, I. Kulaga, and L. Koszalka
S ta rt o f
m u ta tio n
Is th e R a n d o m ly
NO YES B A G C E F H D
c h ro m o s o m e s e le c t 2 g e n e s
im p ro v e d ? 1 p q 8
p and q
A p p ly 2 o p t E xchange B A F C E G H D
a lg o rith m genes 1 p q 8
E nd of
m u ta tio n
4 Investigations
A special experimentation system in Java was created (available on Website [12]).
Apart from the opportunity of random generating cities, the system allows also
loading common TSP problems from TSPLIB [7].
Experiment 1. In Table 1 the average results obtained by 2optGA and 2optGAM for
a randomly generated problem (200 cities) in a specified time horizon are shown. The
size of population was set to 200. Mutation and crossover probabilities were equal to
suitable 0.3 and 0.2, respectively. The experiment was repeated 10 times.
Table 1. The results of algorithm comparison for randomly generated TSP of 200 cities
Table 2. The results of comparison of 2optGA and 2optGAM algorithms for common TSP
problems (the column optimal* contains lengths of optimal routes from TSPLIB)
2optGA 2optGAM
TSPLIB optimal* diff diff
avg c t [s] avg #G t [s]
[%] [%]
att48 33523 33523 0 129 0,8 33523 0 15 0,2
berlin52 7544 7544 0 77 0,6 7544 0 7 0,1
Kroa100 21285 21285 0 122 2,9 21285 0 33 1,4
Kroc100 20750 20750 0 167 3,8 20750 0 52 2,1
eil101 642 645 0,46 3942 68,7 642 0 103 3,6
lin105 14382 14382 0 93 2,6 14382 0 41 1,8
A280 2586 2624 1,46 1133 213,3 2603 0,65 712 179,4
Pr1002 259066 275723 6,42 824 2961,1 264675 2,16 1081 6001,9
The solutions returned by 2optGAM are better in both cases of running time.
Experiment 2. Table 2 presents solutions (the averaged lenghts of routes (avg) and
parameters e.g. the total number of generations (#G), time duration t[s]) obtained for
problems taken from TSPLIB. Algorithms were stopped after finding the optimal
solution or if current best solution was not changed over a specified amount of time.
In most cases, both algorithms produced optimal solutions, however, it took less
time for 2optGAM algorithm. Moreover, in more complex problems (a280, eil101)
the 2optGAM returned not optimal solutions but better than 2optGA improvements.
It may justify the conclusion that the 2optGAM algorithm is very promising.
References
1. Abudiab, M., Starek M., Lumampao R. and Nguyen A.: Utilization of GIS and Graph
Theory for Determination of Optimal Mailing Route. JCSC 4 (2004) 273-278
2. Bryant K., Benjamin A.,: Genetic Algorithms and the Travelling Salesman Problem.
Working Paper, Departament of Mathematics, Harvey Mudd College (2000)
3. Cotta C., Adlana J.F., Troya J.M.: Hybridizing Genetic Algoriths with Branch and Bound
Techniques. Artificial Neural Nets and Genetic Alg., Springer-Verlag (1995) 278-280
4. Kulaga I., Pozniak-Koszalka I.: Gisness System – Cartographic Modelling to Support
Business Decision-Making. Proc. of 4th Polish-British Workshop, Ladek (2004) 93-102
5. Harmon J.E., Anderson S.J.: The Design and Implementation of Geographic Information
Systems. John Wiley, New York (2003)
6. Pozniak-Koszalka I.,: Relational Data Bases in Sybase Environment. Modelling,
Designing, Applications. WPWR, Wroclaw (2004)
7. Reinelt G.: TSPLIB – A Travelling Salesman Problem Library. ORSA Journal on
Computing, 3 (1991) 376-384
8. Reinelt G.: The Travelling Salesman: Computational Solutions for TSP Applications.
Springer-Verlag (1994)
9. Sengoku H.,Yoshihara I.: TSP Solver Using GA on Java. Proc. of AROB (1998) 283-288
10. Shaw S.: Transport Geography on the Web. http://people.hofstra.edu/geotrans/ (2005)
11. website: http://www.zlote.jabluszko.net/tsp/
A DNS Based New Route Optimization Scheme
with Fast Neighbor Discovery in Mobile IPv6
Networks
We offer a fast neighbor discovery scheme to quickly process the DAD procedure
in MIPv6. Our proposed scheme uses the look up algorithm in the modified
neighbor cache of a new access router [5]. The DAD using lookup algorithm
consumes an extremely short amount of time, typically a few micro second units,
such as Longest Prefix Matching speeds in routing table. In Patricia Trie case,
since lookup requires memory access of 48 times in worst case, the number of
lookup value is 48. Hence, Lookup delay is 5.28 μ sec in worst case.
In the conventional DNS-based Mobile IP scheme the DNS update problem
was resolved using the smooth handoff scheme. However, it resulted in data loss
because it could not update quickly in an environment where the MN moves
frequently. To solve this problem we can use local mobility management. When
an MN moves to a foreign network the MN acquires a CoA through a router
advertisement message from the new access router (AR). By constructing routers
hierarchically using Mobility Anchor Point (MAP), we can minimize the delay
produced when an MN updates its CoA in the DNS Resource Record.
As new system architecture could be cost prohibitive, we present a way to
expand DNS. We use inverse domain which changes the name into the address
and maps it. First, we separate a ufl.edu name server to act as a domain name
caching system and an inverse domain system. For example, when an FTP server
receives a packet addressed to 128.227.120.42 from an FTP client, the FTP server
must check whether this FTP client is identified or not. The FTP server could
reference the resource data files which have identified client lists; however, these
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 590–593, 2006.
c Springer-Verlag Berlin Heidelberg 2006
A DNS Based New Route Optimization Scheme 591
files include domain names only, so the FTP server has a “RESOLVER” which
is the DNS client. The RESOLVER sends an inverse query asking the name of
the FTP client to the DNS server and inverse domain system. Through this, the
FTP server acquires the domain name of the FTP client. To support the domain
name we modified the inverse domain to find out the MN’s position in the DNS.
The servers conducting an inverse domain have hierarchical levels. If an MN
receives a new CoA, the MN sends the home address CoA and update request
(registration) message including lifetime to the DNS. Then the DNS confirms the
MN’s home address and the DNS Resource Record. If a mobility biding for the
MN did not exist, the DNS would either create a new resource record or update
the existing one. If the CoA was found, only the lifetime would be updated.
However, if the CoA was not found the existent CoA would be substituted and
the lifetime would be updated as well.
In this section we analyze the delay required for the DNS name resolution. We
separate he signaling procedure for packet delivery in two cases: registration
procedure and routing procedure;
The total routing delay for standard Mobile IPv6 and proposed EDNS-MIPv6
are presented as TRouting−MIP v6 = tCN −HA + tHA−N CoA + tBU −CN + tP acket
and TRouting−EDN S = tDN S−UP + tDN S + tP acket respectively.
Therefore, the total signaling delay (TSignaling ) for packet delivery between
the CN and an MN is the sum of the registration delay (TReg ) and the rout-
ing delay (TRouting ). First of all, we assume that the mobile node receives the
IP address without a domain name. With “TSignaling−MIP v6 = TReg−MIP v6 +
TRouting−MIP v6 ” and “TSignaling−EDN S = TReg−EDN S + TRouting−EDN S ”, we
can find the delay difference, TDif f , between these two mechanisms.
TDif f = {TSignaling−MIP v6 − TSignaling−EDN S } = (tDAD − tLU + tHA−N CoA
+ tCN −HA + tBU −CN − tDN S−UP − tDN S (1)
From Eq.1, TDif f is always a nonnegative integer because “tCN −HA +tBU −CN ”
is greater than or equal to “tDN S−UP + tDN S ” in wireline-network environment.
Also, since tDAD is larger than tLU , when an MN moves to a new network area
the value of TDif f is always positive. What is inferred from “TRouting−MIP v6 ” is
that the important factor affecting the delay is the packet delivery time during tun-
neling from the HA and the MN’s new CoA, tHA−N CoA . If this tunneled packet
delivery time is long, the total delay increases. Therefore, packets sent from the
CN are delivered via the HA and tunneled to the MN until the CN should receives
a binding update message from the MN. Thus, by using the EDNS-MIPv6 scheme
we reduce the total signaling delay and remove the triangle routing problem which
causes degradation of network performance.
3 Conclusion
In this paper, we have proposed a new route optimization scheme (EDNS-MIPv6)
for MIPv6 using the enhanced Domain Name System (DNS) together with fast
neighbor discovery and local mobility management scheme to reduce the signal-
ing delay using lookup algorithm. EDNS-MIPv6 alleviates the triangle routing
problem and reduces both the signaling delay to the home network, and the
home agent’s processing load.
References
1. C. Perkins, K-Y. Wang. “Optimized smooth handoffs in Mobile IP”, Proceedings of
IEEE Symposium on Computers and Communications, Egypt, July 1999.
2. M.Conti, E.Gregori, S.Martelli, “DNS-based architectures for an efficient manage-
ment of Mobile Users in Internet ”, IEEE, 2001
3. Byungjoo Park, Sunguk Lee, Haniph Latchman, “Performance Analysis of Enhanced
Mobile IPv6 with Fast Handover over End to End TCP”, in. Proc. of IEEE Wireless
Com-munications and Networking Conference (WCNC), April 3-6, 2006.
Performance Analysis of Group Handoff in
Multihop Mesh Relay System
1 Introduction
The multihop mesh relay(MMR) is a promising solution to expand coverage,
enhance throughput and system capacity to wireless broadband multimedia ser-
vices such as IEEE 802.16 systems. The gains in coverage and throughput can
be leveraged to reduce total deployment cost for a given system performance re-
quirement and thereby improve the economic viability of IEEE 802.16 systems.
In this system, all resources such as Channel Elements(CE) and call process-
ing hardwares are located in the CS. In the CS, call processing hardwares and
resources are shared among multiple relays and this can improve a trunking ef-
ficiency. However this makes the increase of handoff in the system. Also there
are many hot-spot cells because of the variety of different moving patterns of
subscribers. To solve these problems, the simulcast technique can be adopted
[1]. In case that low traffic load arises, multiple relays can be grouped to share
resources. All relays within a group broadcast and receive signals as if they are
in the same cell. This grouping of relays can be dynamically changed accord-
ing to a traffic distribution. By rearranging the simulcast group dynamically,
the system can reduce handoff rates and protect the outbreak of hot-spot cell.
When groups are reorganized by dynamic group simulcasting, a relay which is
located in a group can be transferred to the other group. In this case, all calls
in that relay must be handoffed simultaneously to the target group. This type
of handoff is named ’Group Handoff’. In this paper, we analyze the performance
of group handoff in MMR system.
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 594–597, 2006.
c Springer-Verlag Berlin Heidelberg 2006
Performance Analysis of Group Handoff in Multihop Mesh Relay System 595
algorithm. In this paper, we analyze using the simplest dynamic grouping algo-
rithm which sets up new grouping due to the number of available channels in
each group. We define that the system consists of two simulcast groups which
are controlled by a CS. Each group has limited number of channel, C, which is
the maximum number of simultaneous calls that can be serviced in the group.
We define that initial number of relays in group 1 and group 2 are N1 and
N2 , respectively. We define that there should be at least two relays in a group
after group handoff. Also, we assume that there is no handoff arrival from the
neighboring system. Dynamic grouping is executed because of nonuniform traffic
distribution of each group due to different call arrival in each group. We assume
that each group has different new call arrival rate and it is changed after new
grouping is established. We suppose that relays in a group have the same new
call arrival rate. Let new call arrival rate in group 1 and group 2 be λn1 and λn2 ,
respectively. Let handoff call arrival rate from neighbor groups be λh and group
handoff generation rate be λg . We assume that call duration time, Tc is exponen-
tially distributed with mean μ−1 c . Let cell dwell time be Td which is exponentially
distributed with mean μ−1 d . We define two threshold values to decide timing of
grouping and select relay adequate to be handoffed: T Hh is upper threshold and
T Hl is lower threshold. We divide total capacity of a group into three levels. If
state of a group is Level 1, only group handoff-out can be generated. If a group
is in state of Level 2, this group can receive handoffed relay from other group
and send a relay to other group. In state of Level 3, only group handoff-in can
occur.
Using birth-death process, we can derive the state transition diagram in view
of total system[2]-[3]. Each state is assigned to s and the state of this process is
defined as s = (i, j, k, l) where i and k are the number of communicating calls
in group 1 and group 2, respectively. Also, j and l are the number of relays in
group 1 and group 2, respectively. There are five sets of state-transition. They
are symbolized by character from A to E. X and X̄ make a reversible pair. We
define the transition to right direction be X1 and the transition to left direction
be X2 if the set of transition is X. A1 and B1 mean that a new call is originated in
group 1 and group 2, respectively. When a relay is handoffed to other group, call
arrival rate of the relay is not changed. So, the total call arrival rate of group 1,
λ1 is calculated as λn1 + (j − N1 ) λNn22 when (j ≥ N1 ), and Nj1 λn1 when (j < N1 ).
Also, the total call arrival rate of group 2, λ2 is given by λn2 + (l − N2 ) λNn11 when
(l ≥ N2 ) and Nl2 λn2 when (l < N2 ). A2 and B2 mean that a call in group 1 and
group 2 are ended, respectively.
D1 is the transition when a relay in group 1 is handoffed out to group 2. The
number of calls in handoffed relay, n is defined as
i i i i i
n = { − x, − (x − 1), · · · , , · · · , + (x − 1), + x }. (1)
j j j j j
There are (2x + 1) cases of group handoff, where x is the variation factor. If
we assume that occurring probability of each value is the same, the transition
rate of group handoff is expressed as 1j · 2x+1
1
· λg . The first term, 1/j means the
596 Y.-u. Chung, Y.-H. Choi, and H. Lee
probability of select one relay in a group. The term, 1/(2x + 1) is the probability
that the selected relay has value of n. We assume that λg has three sorts of
values, λg1 , λg2 , and λg3 according to events. λg1 has the largest value. If “Level
1 ⇒ Level 3” event occurs, λg1 is adopted. When “Level 1 ⇒ Level 2” and “Level
2 ⇒ Level 2” events occur, λg2 is adopted. In the event of “Level 2 ⇒ Level 3”,
λg3 is adopted.
E1 means the case that a group handoff is occurred from group 2 to group 1.
From the same way of the case of D1, Each n and transition rate is expressed as
k k k k k
n = { − x, − (x − 1), · · · , , · · · , + (x − 1), + x } (2)
l l l l l
1 1
transition rate = · · λg . (3)
l 2x + 1
We evaluate the performance of proposed scheme in view of blocking and
handoff call dropping probability. Let P (s) be the steady-state probability of
state, s. A new call is blocked when all channels in the group are occupied. This
case occurs when i ≥ C or k ≥ C. And the blocking probability is given by
2 −1 N1 +N
N1 +N 2 −1
C
PB = P (i, j, C, l)|j+l=N1 +N2 (4)
j=1 l=1 i=0
2 −1 N1 +N
N1 +N 2 −1
C
+ P (C, j, k, l)|j+l=N1 +N2 − P (C, j, C, l)|j+l=N1 +N2 .
j=1 l=1 k=0
As we assume that there is no queue and reservation channel for handoff call,
blocking probability and handoff call dropping probability have the same value.
3 Numerical Results
We assume that x has the value of 1. And the number of calls in the relay, n
have three types of value. We investigate several numerical examples in case of
μc = 0.01, μd = 0.03, C = 20, and N1 = N2 = 4. To evaluate the performance
of group handoff in this system, We consider two conditions. First, we generate
numerical results with changing the disparity of call arrival rate in each group.
We assume that λg1 = 0.9, λg2 = 0.5, λg3 = 0.1, T Hh = 16, and T Hl = 4. We
compare performance in case of executing dynamic grouping and fixed grouping.
The result is shown in Fig. 1. In Fig. 1, D and F means the case using dynamic
grouping and fixed grouping, respectively. a : b means the fraction of new call
arrival rate in group 1 and group 2. From the result, blocking probability and
handoff call dropping probability of dynamic grouping has similar value in case
that the difference of call arrival rate in each group is changed. But, in case of
fixed grouping, we could see that the larger the difference of call arrival rate in
each group, the higher the blocking probability and the handoff call dropping
probability.
Performance Analysis of Group Handoff in Multihop Mesh Relay System 597
The second condition is the case that the values of λg are changed. In this
case, we assume that T Hh = 16, T Hl = 4, and the fraction of new call arrival
rate in group 1 and group 2 is 2 : 1. The result is also shown in Fig. 1. In
Fig. 1, λg1 = 0.9 means that (λg1 , λg2 , λg3 ) = (0.9, 0.5, 0.1). λg1 = 0.5 means
that (λg1 , λg2 , λg3 ) = (0.5, 0.3, 0.1). And, λg1 = 0.3 means that (λg1 , λg2 , λg3 ) =
(0.3, 0.2, 0.1). The results show that the higher the value of λg , the lower the
blocking probability and the handoff call dropping probability. That’s because
high value of λg can generate group handoff when it is needed. So, the blocking
probability is diminished when the value of λg is high.
4 Conclusions
The MMR system was proposed to to expand coverage, enhance throughput and
system capacity. In this system, simulcast technique using dynamic grouping can
rearrange traffic load easily. Group handoff is generated by dynamic grouping.
In this paper, we modeled group handoff using the Markov chain and performed
system analysis. From numerical results, we can see that the larger the difference
of call arrival rate in each group, the better the performance of dynamic grouping.
Also, numerical result shows that dynamic grouping provides better performance
than fixed grouping in simulcast technique.
References
1. S. Ariyavisitakul, T. E. Darcice, L. J. Greenstein, M. R. Philips and N. K. Shankara-
narayanan, ”Performance of Simucast Wireless Techniques for Personal Communi-
cation Systems,” IEEE JSAC, Vol. 1, No. 4, pp.632-643, May 1996.
2. S. S. Rappaport, ”The Multiple-Call Hand-Off Problem in High-Capacity Cellular
Communication Systems,” IEEE Trans. on Vehicular Technology, Vol. 40, No. 3,
pp.546-557, Aug. 1991
3. S. -L. Su, J. -Y. Chen and J. -H. Huang, ”Performance Analysis of Soft Handoff in
CDMA Cellular Networks,” IEEE JSAC, Vol. 14, No. 9, pp.1762-1769, Dec. 1996.
DSMRouter: A DiffServ-Based Multicast
Router
Yong Jiang
1 Introduction
The Internet Engineering Task Force (IETF) has proposed the Differentiated
Service (DiffServ) architecture [1][2] to solve the Quality of Service (QoS) for
different users and numerous types of traffic.
As we know, in an MPEG-coded video system, packet loss causes unexpected
degradation of the quality of the received video, 3% packet loss in an MPEG-
coded bit stream can be translated into a 30% frame error rate [3]. Shin et al.
proposed a content-based packet video forwarding mechanism on the DiffServ
network [4][5]. In [6], Striegel et al. proposed an architecture for DiffServ-based
multicast that achieves the benefits of traditional IP multicast but does not
require per group state information in the core routers.
In this paper, we realized a DiffServ-based Multicast Router (DSMRouter)
based QoS model, for multicasting video transmission in the DiffServ network.
The layered method [7] is used to transmit video streams, including I/P/B
frames. One video stream is separated into several types of media flows us-
ing so-called layered media. When layered media are transmitted with different
This research was sponsored by NSFC (No. 60503053) and GDNSF (No. 034308),
and Development Plan of the State Key Fundamental Research (973) (No.
2003CB314805).
Y.-T. Kim and M. Takano (Eds.): APNOMS 2006, LNCS 4238, pp. 598–601, 2006.
c Springer-Verlag Berlin Heidelberg 2006
DSMRouter: A DiffServ-Based Multicast Router 599
3 System Architecture
A DiffServ-based Multicast Router (DSMRouter) can be deployed at the edge of
the DiffServ domain. DSMRouter can classify, policy and shape the traffic that
transmit through it, just as one general DiffServ edge node. DSMRouter can
provide different QoS guarantees according to the Per Hop Behavior (PHB).
A traffic sending mechanism under the PFS strategy, for streaming video,
operates to enable transmitted streams between different DiffServ service classes
to guarantee different QoS levels of the DiffServ network. The video stream is
split into four classes-I frame, P frame, B frame and audio streams. Different
classes have different delay bounds, loss bounds and bandwidth requirements.
The delay, the packet loss and the bandwidth requirements of different frame
streams can be analyzed to mark their DSCPs (Differentiated Service code-point)
with different QoSs. In the DiffServ network, the layered video streams are sent
by multicast. The frame streams are sent in the same multicast session with
different port numbers and DSCP values.
4 Performance Evaluation
In this section, we have implemented our traffic sending mechanism under the
PFS strategy in the DiffServ network, and evaluated the performance of the
layered media scheme that is used in the DSMRouter.
/RVVUDWH
7LPHVHF
Fig. 1 shows the loss rate distribution of I/P/B layered media streams. Clearly,
PFS strategy achieves the loss rate proportional differentiation well. In 0 to 400
sec, media flows are transmitted through the DSMRouter with the PFS traffic
sending mechanism. Under light load situation, the loss rate of I/P/B indicates
that almost no data are lost with such a bandwidth reservation. While after 400
sec, the addition of the heavy load flow causes the loss rate of I/P/B to start
to burst, but the layered media with higher priority have less packet loss and so
exhibit improved presentation to end users when the traffic is heavy.
DSMRouter: A DiffServ-Based Multicast Router 601
5 Conclusion
This paper has proposed a DiffServ-based media gateway called the DiffServ-
based Multicast Router (DSMRouter) to guarantee QoS and scalability for mul-
timedia presentation. In the reserving bandwidth situation, the approximate
amounts of reserved bandwidth are computed for different service queues by
applying the PFS strategy. With regard to the dynamic bandwidth adjustment
situation, DSMRouter can be con-figured to serve different classes of layered
media with various resource reservations based on the PFS strategy. The results
of the performance analysis indicate that the system can detect the congestion
of the network and react by immediately implement-ing appropriate procedures.
References
1. S. Blake, D. Black, M. Carlson, E. Davies, Z.Wang, andW.Weiss, ”An Architecture
for Differentiated Services,” RFC 2475, IETF, Dec. 1998.
2. P. Trimintzios, I. Andrikopoulos, G. Pavlou, C. F. Cavalcanti, D. Goderis, Y.
T’Joens, P. Georgatsos, L. Georgiadis, D. Griffin, C. Jacquenet, R. Egan, and
G. Memenios, ”An archi-tectural framework for providing QoS in IP differenti-
ated services networks,” in Proc. 7th IFIP/IEEE Int. Symp. Integrated Network
Management (IM 2001), May 2001, pp. 17-34.
3. J. Boyce and R. Gaglianello, ”Packet loss effects on MPEG video sent over the
public internet,” in Proc. 6th ACM Int. Conf. Multimedia, Sep. 1998, pp. 181-190.
4. J. Shin, J. W. Kim, and C.-C. J. Kuo, ”Content-based packet video forwarding
mechanism in differentiated service networks,” in Proc. Int. Packet Video Work-
shop, May 2000.
5. J. Shin, J. W. Kim, and C.-C. J. Kuo, ”Quality of service mapping mechanism for
packet video in differentiated services network,” IEEE Trans. Multimedia, vol. 3,
no. 2, pp. 219-231, Jun. 2001.
6. A.Striegel, G. Manimaran, ”A Scalable Approach for DiffServ Multicasting”, Proc.
of Inter-national Conference on Communications, Helsinki, Finland, June 2001.
7. M. Kawada, K. Nakauchi, H. Morikawa, and T. Aoyama, ”Multiple streams con-
troller for layered multicast,” in Proc. IEEE Int. Conf. Communications, vol. 1,
1999, pp. 65-68.
8. Dovrolis C, Stiliadis D. Relative Differentiated Services in the Internet: Issues
and Mecha-nisms. In: Proceedings of ACM SIGMETRICS’99, Atlata , May 1999.
pp204-205
9. Y. Jiang, C. Lin, and J. Wu. Integrated performance evaluating criteria for network
traffic control. In: Proceedings of IEEE symposium on Computers and Communi-
cations 2001, IEEE Communications Society Press, Tunisia, July 2001.
10. Y. Jiang, J. Wu, The proportional fairness scheduling algorithm on multi-classes,
Science in China Series F, Vol.46 No.3, June 2003, p 161-174.
Author Index