tail : .15.. é. H... B,- ’PJWJ: £th a .. . 510317....- .1} new in a. .1‘ 2.9.4. , a... fax. ‘21-: 3:59.: 24.1.} 1...?“ x . Linn-V“ , ..‘):~!.(I w . . .. (b. iafihfi. valetii :umWL 4 s. .¥2.§:§ ‘u'lt‘§l‘“fl! . $.38: 1113533 LIBRARY Michigan State University x This is to certify that the dissertation entitled OFF-NETWORK CONTROL PROCESSING FOR SCALABLE ROUTING IN VERY LARGE SENSOR NETWORKS presented by Tao Wu has been accepted towards fulfillment of the requirements for the PhD. degree in Electrical Engineerim omitlixw. Major Professor's Signature :0 /e 0/0 a l 7 Date MSU is an Affirmative Action/Equal Opportunity Employer PLACE IN RETURN BOX to remove this checkout from your record. TO AVOID FINES return on or before date due. MAY BE RECALLED with earlier due date if requested. DAIEDUE DAIEDUE DAIEDUE 5/08 K:lProj/Acc&Pres/ClRCIDateDueindd OFF-NETWORK CONTROL PROCESSING FOR SCALABLE ROUTING IN VERY LARGE SENSOR NETWORKS By Tao Wu A DISSERTATION Submitted to Michigan State University In partial fiilfillment of the requirements For the degree of DOCTOR OF PHILOSOPHY Electrical Engineering 2008 ABSTRACT OFF-NETWORK CONTROL PROCESSING FOR SCALABLE ROUTING IN VERY LARGE SENSOR NETWORKS By Tao Wu Although wireless sensor networks have demonstrated a promising future, there are still a significant number of challenges to be overcome before they become reality. The prime research issues involve fault tolerance, network scalability, cost, hardware design, topology management, and power consumption. In this thesis we investigate two problems concerning wireless sensor networks: (A) energy efficient medium access control and (B) scalable routing architecture for very large wireless sensor networks. Regarding problem A, we propose a Self Reorganizing Slot Allocation (SRSA) mechanism for a TDMA-based medium access control (MAC) protocol in wireless sensor networks. With SRSA, a node can achieve significant energy savings by remaining active only during allocated slots for transmissions and receptions. In multi-cluster networks, it is often necessary for nodes to use either CDMA or FDMA for preventing TDMA slots across neighboring clusters from interfering with each other. Using FDMA or CDMA incurs costs in terms of spectrum usage as well as hardware complexity. Our goal is to provide an alternative design that can reduce inter-cluster TDMA interference without having to use CDMA or FDMA. The primary contribution of our work is to demonstrate that with adaptive slot allocation algorithms, it is possible to minimize such interference under low traffic loading conditions. The second contribution is to design a feedback-based adaptive allocation protocol that can minimize those interferences without relying on any global synchronization mechanisms. We present the design of SRSA and provide a detailed simulation-based characterization of the protocol. In the second problem, we motivate an architectural solution to address the problem of scalable routing in very large sensor networks. The control complexities of the existing sensor routing protocols, both node-centric and data-centric, do not scale well for large networks with potentially hundreds of thousands of embedded sensor devices. We develop a routing solution Off-Network Control Processing (ONCP) that achieves control scalability in large sensor networks by shifting a certain amount of routing functions to an “off-network” server. A tiered hybrid routing approach, consisting of “coarse grain” server-based global routing, and distributed “fine grain” local routing is proposed for achieving scalability by avoiding network-wide control message dissemination. We present the ONCP architectural concepts and analytically characterize its performance in relations to both flat and hierarchical routing architectures. Through an nsZ-based simulation model, the experimental results indicate that for large sensor networks with realistic data models, the packet dr0p, A latency and energy performance of ONCP can be significantly better than those for a flat sensor routing protocol such as Directed Diffusion and hieratical cluster-based protocol like CBRP. We also incorporate the sink mobility into routing architecture, and investigate its impacts on the ONCP, Directed Diffusion, and CBRP routing protocols. To my wife, Lan Yang, my father; Daxin Wu and in memory of my mother: Yzzhi Zhou iv ACKNOWLEGDEMENTS I would like to take this opportunity to acknowledge all the help and support I have received over the years in Michigan State University. No amount of words can do this adequately. First of all, I am extremely fortunate to have Dr. Subir Biswas as my thesis advisor, who committed years in training and helping me to be a qualified researcher. Under his excellent guidance, I have learnt to be a researcher with a lifelong interest in academic area. He offered valuable insights and suggestions in guiding my research directions. He read my paper and dissertation drafts and edited my grammar errors without a complaint. I am also grateful for his excellent advices regarding to presentations, writings and many other non-technical endeavors. It is an honor and privilege to be his student and to work with him. I would be delighted if I can still keep his high standards after graduation. I am also very thankful for Dr. Biswas’s family, who welcomed me warmly. I wish to thank Dr. Stuart H. Gage, a member of my Ph.D. committee. As a biological professor, Dr. Gage gave the knowledge and insight I needed to apply the research principles to the real world problems. He taught me to think about research problems with broad perspectives, and encouraged me to study practical problems at the same time. He provided many suggestions in designing experiments. I also appreciate the help from other members of my Ph. D. committee: Dr. Eric Goodman, Dr. Peixin Zhong, and Dr. Jian Ren. They encouraged me in the process of this dissertation and offered invaluable feedbacks. I could not have accomplished anything without the support from all my fellow students and lab-mates: J ayanthi Rao, Fan Yu, Darshak Thakore, Rohit Naik, Nagendra Singh, Anthony Plummer, Howard Hardiman, Muhannad Quwaider, Sonny Gupta, Raymond Tatchikou, etc. I thank all of them for many useful discussions, help and friendship. I thank all the professors at Michigan State University whom I have taken courses from and interacted with: Professor Matt W. Mutka, Eric Tomg, Li Xiao, John R. Deller, Oweiss Karim, and Yimin Xiao. I especially thank Professor Yimin Xiao for demonstrating that statistics is interesting and fun. Last, and most importantly, I would like to thank my family. They are the reasons for who I am, and they are always encouraging and supporting me in all times. I would like to thank my wife Lan Yang for her support and love during the past few years. Her understanding and encouragement is very important for my dissertation. My parents, Daxin Wu and erhi Zhou, receive my deepest gratitude for their many years of support during my undergraduate and graduate studies. I wish I could enumerate all the people I need to give thanks, but I am sure I missed many of them on this short acknowledgment. I am grateful for each of them and wish I could pass their spirits and kindness to other people in the future. This work was supported by an NSF grant (SCI-O43 8271) on Sensor Network Protocols. And most of simulations in this thesis were performed in the cluster computers of the High Performance Computing Center (HPCC) in Michigan State University. vi TABLE OF CONTENTS LIST OF TABLES ................................................... x LIST OF FIGURES ................................................... xi 1. Introduction 1 1.1. Wireless Sensor Networks ..................................... l 1.2. Wireless Sensor Network Applications ............................ 6 1.2.1. Environmental Data Collection .............................. 6 1.2.2. Event Based Monitoring ................................... 8 1.2.3. Object Tracking ......................................... 10 1 .2.4. Other Applications ....................................... 1 1 1.3. Research Issues on Energy Efficiency and Network Scalability ....... 12 1.3.1. Energy Efficient Medium Access Control ...................... 13 1.3.2. Scalable Routing for Very Large Sensor Network ............... 14 2. A Self Reorganizing Slot Allocation Protocol for Multi-cluster Sensor Networks 18 2.1. Introduction and Motivation .................................... 18 2.2. Related Work .............................................. 21 2.3. Self-Reorganizing Slot Allocation Protocol (SRSA) ................ 24 2.3.1. Network Model and Assumptions ........................... 24 2.3.2. SRSA Protocol Overview ................................... 26 2.3.3. Inter-cluster Allocation Dependency .......................... 28 2.3.4. Frame Scaling for Collision Reduction ....................... 30 2.3.5. Allocation Reorganization Algorithm ........................ 31 2.3.6. Dimensioning Frame Scaling Factor ......................... 35 2.3.7. Protocol Convergence ..................................... 38 2.4. Performance Evaluation ...................................... 39 2.4. 1. Experimental Parameters .................................. 39 2.4.2. Simulation Results ....................................... 40 2.5. Summary ................................................. 53 3. Scalable Routing for Very Large Sensor Network 55 3.1. Related Work on Sensor Routing Protocols ....................... 55 3.1.1. Flat Sensor Routing Protocols ............................ 57 3.1.2. Hierarchical Sensor Routing Protocols ..................... 58 3.1.3. Geographic Sensor Routing Protocols ...................... 59 3.2. Motivation and Problem Description ............................ ‘ .60 3.3. Network And Application Model ............................... 61 3.4. Off-Network Control Processing (ONCP) Routing Architecture ...... 62 3.4.1. Region Management ................................... 64 3.4.2. Cluster Organization ................................... 66 vii 3.4.3. Status Updates ........................................ 68 3.4.4. Global Routing ....................................... 69 3.4.5. Command & Global Routing Download ................... 72 3.4.6. In-network Local Routing .............................. 73 4. Protocol Overhead Analysis 76 4.1. Notations and Definitions ........................................ 76 4.2. Overhead Analysis of Flat Protocols. ........................... 77 4.3. Overhead Analysis of Hierarchical Cluster-based Protocols .......... 78 4.4. Overhead Analysis of ONCP protocol ............................ 79 4.5. Numerical Results ........................................... 81 Il.t5. Eitrrrirriztrjr ................................................. 2355 5. Simulation Models and Experimental Setup 86 5.1. NS-2 Simulation Models ..................................... 86 5.1.1. Channel Propagation Model ............................. 86 5.1.2. Radio Energy Model .................................. 88 5.1.3. Application Data Model ................................ 89 5.2. ONCP Routing Implementation ................................ 91 5.2.1. ONCP Agent Structure ................................ 92 5.2.2. Cluster Organization .................................. 94 5.2.3. Status Updates ....................................... 97 5.2.4. Global Routing and Command Forwarding ................. 98 5.2.5. Local Routing Mechanism ............................. 100 6. Experimental Performance Evaluation of ONCP 103 6. 1 . Experimental Environment ................................... 103 6.2. Results and Analysis ........................................ 105 6.2.1. The Scenarios with Single Base Station .......................... 105 6.2.2. The Scenarios with Multiple Base Stations ........................ 116 6.2.3. Impacts of Region Status Updates ................................ 131 6.2.4. Impacts of Network Size ....................................... 133 t5.22.fi. Eitrrrrrriztryi ................................................... 12315 7. Off-Network Control Processing with Sink Mobility 137 7.1. Introduction and Motivation .................................... 137 7.2. Related Work ............................................... 140 7.3. ONCP-M: Off-Network Control Processing with Sink Mobility ...... 142 7.3.1. Network and Data Model for ONCP with Sink Mobility ...... 142 7.3.2. Extension of ONCP Framework .......................... 144 7.3.3. Sink Relocation Algorithm .............................. 146 7.4. Simulation Results .......................................... 148 17.15. ESirrrirrrzrrjyr .................................................. 1.6577 viii 8. Summary and Future Work 168 8.1. Summary .................................................. 168 8.2. Future Work ................................................ 170 APPENDIX A Functional Validation of ONCP through Software Emulation and Hardware Prototype Implementation 172 A]. Introduction ................................................ 172 A2. TinyOS Operating System ................................... 173 A3. ONCP Protocol Stack Implementation .......................... 175 AA. Software Emulation Results .................................. 181 A5. Prototype Hardware Implementation ............................ 184 REFERENCES ...................................................... 187 LIST OF TABLES Table 2.1: Allocation possibilities for the network in Fig.2.2 ............................. 30 Table 2.2: Overall performance summary ................................................... 53 Table 5.1: Radio Parameters .................................................................. 89 LIST OF FIGURES F ig.2. 1: Target sensor network and data flow model ....................................... 25 Fig.2.2: TDMA Collisions across overlapping clusters .................................... 28 Fig.2.3: Example slot allocations for network in Fig.2.2 ................................... 29 F ig.2.4: Use of frame scaling for collision reduction ....................................... 31 Fig.2.5: Pseudo code of SRSA protocol logic at each cluster head ...................... 34 Fig.2.6: Parameters for Frame Scaling Factor dimensioning .............................. 36 Fig.2.7: Comparison between SRSA and ideal allocation .................................. 37 Fig.2.8: Model to represent system state with SRSA ........................................ 39 Fig.2.9: Throughput as a function of load ..................................................... 41 Fig.2. 10: CS and I-IN collisions as a function of load ....................................... 42 Fig.2.11: Protocol convergence in SRSA ..................................................... 44 Fig.2.12: Protocol convergence for varying network size .................................. 45 F ig.2. 13: Spatial distribution of packet drops ................................................ 46 Fig.2.14: Packet delivery latency as a function of load .................................... 47 Fig.2. 15: Energy expenditure as node uptime ................................................ 48 Fig.2.16: Collisions for different cluster overlapping ....................................... 49 Fig.2. 17: Throughput with varying cluster overlapping .................................... 50 Fig.2.18: Effects of flame scaling on collisions in SRSA .................................. 51 Fig.2.19: Effects of mobility on Collisions in SRSA ....................................... 52 Fig.3.1: Network and application model ...................................................... 62 Fig.3.2: Tiered concept of ONCP routing .................................................... 63 Fig.3.3: Example region-to-region connectivity metric .................................... 65 Fig.3.4: Pseudo code of ONCP global path computation ................................... 71 Fig.3.5: Pseudo code of command & global routing forwarding .......................... 72 xi Fig.3.6: Local routing in ONCP architecture ................................................ 73 Fig.4.1: Impact on network size on relative transmission ratio ............................ 82 Fig.4.2: Impacts of data per request (D) on transmission ratio ............................. 83 Fig.4.3: Impacts of region size on the transmissions ratio ................................... 84 Fig.5. 1: Internal blocks of mobile node ...................................................... 92 Fig.5 .2: ONCP agent diagram .................................................................. 93 F ig.5.3: Format of neighbor hello message ................................................... 95 F ig.5.4: Pseudo code for cluster organization algorithm .................................... 96 Fig.5.5: Format of status update message .................................................... 98 Fig.5.6: Format of command message ...................................................... 100 F ig.5.7: Format of local route request message ............................................. 101 Fig.5.8: Format of local route reply message ............................................... 101 F ig.6. 1: Timing diagram of sensing task requests from user clients ..................... 103 Fig.6.2: The diagram of sensing target area and number of nodes per target area. . 104 Fig.6.3: Normalized protocol overhead at rates of 0.5 and l packet/sec ............... 106 Fig.6.4: Normalized protocol overhead at rates of 2 and 4 packet/sec .................. 106 Fig.6.5: Expected packet delivery ratio rates of 0.5 and 1 packet/sec ................... 107 F ig.6.6: Expected packet delivery ratio rates of 2 and 4 packet/sec ..................... 108 Fig.6.7: End to end packet delivery latency at rates of 0.5 and 1 packet/sec ........... 109 Fig.6.8: End to end packet delivery latency at rates of 2 and 4 packet/sec .............. 110 Fig.6.9: Average hop counts at rates of 0.5 and 1 packet/sec ............................. 111 F ig.6. 10: Average hop counts at rates of 2 and 4 packet/sec .............................. 112 Fig.6.11: Average dissipated energy at rates of 0.5 and 1 packet/sec .................... 113 Fig.6. 12: Average dissipated energy at rates of 2 and 4 packet/sec ...................... 114 Fig.6.13: Communication energy consumption at rates of 0.5 and l packet/sec ....... 115 xii F ig.6. 14: Communication energy consumption at rates of 2 and 4 packet/sec ......... 116 Fig.6.15: Normalized protocol overhead at rates of 0.5 and 1 packet/sec (2BS) ...... 118 Fig.6. l6: Normalized protocol overhead at rates of 2 and 4 packet/sec (2BS) ......... 118 Fig.6.17: Normalized protocol overhead at rates of 0.5 and l packet/sec (4BS) ...... 119 Fig.6.18: Normalized protocol overhead at rates of 2 and 4 packet/sec (4B 8) ......... 119 Fig.6.19: Expected packet delivery ratio at rates of 0.5 and 1 packet/sec (ZBS) ....... 121 F ig.6.20: Expected packet delivery ratio at rates of 2 and 4 packet/sec (2BS) ......... 121 Fig.6.21: Expected packet delivery ratio at rates of 0.5 and 1 packet/sec (4BS) ....... 122 Fig.6.22: Expected packet delivery ratio at rates of 2 and 4 packet/sec (43 S) ......... 122 Fig.6.23: End to end packet delivery latency at rates of 0.5 and 1 packet/sec (23S)...123 F ig.6.24: End to end packet delivery latency at rates of 2 and 4 packet/sec (2BS).... 124 Fig.6.25: End to end packet delivery latency at rates of 0.5 and l packet/sec (438)... 125 Fig.6.26: End to end packet delivery latency at rates of 2 and 4 packet/sec (43 S). . 125 Fig.6.27: Average hop counts at rates of 0.5 and 1 packet/sec (2BS) .................... 126 Fig.6.28: Average hop counts at rates of 2 and 4 packet/sec (2BS) ...................... 126 Fig.6.29: Average hop counts at rates of 0.5 and l packet/sec (4BS) ................... 127 Fig.6.30: Average hop counts at rates of 2 and 4 packet/sec (4BS) ...................... 127 Fig.6.3 1: Average dissipated energy at rates of 0.5 and 1 packet/sec (2BS) ............ 129 Fig.6.32: Average dissipated energy at rates of 2 and 4 packet/sec (2BS) ............... 129 Fig.6.33: Average dissipated energy at rates of 0.5 and 1 packet/sec (4BS) ............ 130 Fig.6.34: Average dissipated energy at rates of 2 and 4 packet/sec (4BS) .............. 130 Fig.6.35: Impacts of node failure on a) delivery performance and b) quality of global routes ............................................................................................ 132 Fig.6.36: Impacts of status update on energy dissipation ................................. 133 Fig.6.37: Expected packet delivery ratio (13:6) ............................................. 134 Fig.6.38: Communication energy consumption (13:6) .................................... 135 xiii F ig.7. 1: Network and Application Model with Mobile Base Stations ................... 143 Fig.7.2: Pseudo code for Joint Sink Relocation Algorithm ............................... 148 F ig.7.3: Normalized protocol overhead ..................................................... 15 1 Fig.7.4: Expected packet delivery ratio ..................................................... 152 Fig.7.5: Normalized energy consumption .................................................. 153 F ig.7.6: End to end packet delivery latency ................................................ 153 Fig.7.7: Average transmission hops ......................................................... 154 F ig.7.8: Energy variance at data rate 1 packet per second ................................. 155 Fig.7.9: Energy variance at data rate 2 packet per second ................................ 155 Fig.7.10: Energy variance at different times (data rate = 4 packets per second and 1 node per target area) ................................................. ' .......................... 1 57 Fig.7.11: Energy variance vs the received packets by the base station (data rate = 4 packets per second and 1 node per target area) ............................................ 158 F ig.7.12: Energy variance reduction (data rate = 4 packets per second and 1 node per target area) ...................................................................................... 15 8 Fig.7.13: Node-energy distribution (data rate = 4 packets per second and 1 node per target area) ...................................................................................... 159 Fig.7. l4.A: Node-energy distribution at 1080 see (Directed Diffusion - M) ......... 161 Fig.7.14.B: Node-energy distribution at 1820 see (Directed Diffusion - M) ......... 161 Fig.7 . 14.C: Node-energy distribution at 2550 sec (Directed Diffusion — M) ......... 162 Fig.7. l4.D: Node-energy distribution at 3260 sec (Directed Diffusion — M) ......... 162 Fig.7 . 1 5.A: Node-energy distribution at 1080 see (CBRP — M) ........................ 163 Fig.7 . 1 SB: Node-energy distribution at 1820 see (CBRP — M) ........................ 163 Fig.7.15.C: Node-energy distribution at 2550 see (CBRP — M) ........................ 164 Fig.7.15.D: Node-energy disuibution at 3260 sec (CBRP — M) ........................ 164 Fig.7. 16.A: Node-energy distribution at 1080 sec (ONCP — M) ........................ 165 Fig.7. 16.B: Node-energy distribution at 1820 see (ONCP — M) ........................ 165 xiv Fig.7.16.C: Node-energy distribution at 2550 see (ONCP — M) ........................ 166 Fig.7.16.D: Node-energy distribution at 3260 see (ONCP — M) ........................ 166 Fig.A. l: ONCP Protocol Stack Component Graph ........................................ 176 Fig.A.2: Expected Delivery Ratio ........................................................... 182 Fig.A.3: Routing Overhead .................................................................. 183 Fig.A.4: Average Transmission Hops ....................................................... 184 F ig.A.5: Sensor Node Mote2dot, Mote2, Imote, TelosB, Cricket (from left to right) ........................................................................................... .185 Fig.A.6: Grid structure of hardware implementation ........................... 186 XV Chapter 1 Introduction 1.1 Wireless Sensor Networks Recent advances in micro-electro-mechanical system (MEMS) and integrated digital electronics make it possible to integrate sensing, computation, communication into single tiny devices. Such devices are generally equipped with the capabilities of data processing and communication. These inexpensive and low power devices can provide high-resolution measurement of physical environment, processes and location information to enable a wide range of applications. While the capabilities of a single device can be often limited, the composition of hundreds or thousands of such devices forms wireless sensor networks (W SN). Such networks promise advantages over traditional sensing methods in many ways. Large scale deployment and close proximity to physical environments not only extend the spatial coverage and increase orders of magnitude measurement granularity, but also provide fault-tolerance and robustness for systems. The major strength of wireless sensor networks results from the ability of large scale node deployment and in-network data processing at very low cost. The network size can be scaled to thousands or even hundreds of thousands of wireless sensor nodes in applications, such as environmental monitor, battlefield surveillance and enemy tracking. By the year 2010, the price of these devices is expected to be less than $1 [1]. At this price range, with a communication range of 50 meters, it costs only 40 thousand dollars to fully monitor a 100 square kilometer habitat area. Also, the deployment costs for wireless sensor networks can be minimal. Unlike traditional wired systems, network operators do not need to deploy thousands of feet of wires to connect the sensor nodes. The network can be easily extended by adding more nodes into the network without additional configuration. And in case of node failures, wireless sensor networks have the ability to self-reorganize and reestablish network topologies. While wireless sensor networks share many commonalities with Ad-hoc networks and wireless mesh networks, there are a number of differences and unique challenges due to their specific design characteristics. Such challenges arise from constrained energy supply, limited bandwidth and processing ability, and ultra large scale of node deployment. The unique characteristics of wireless sensor networks are highlighted as follows. Applica_tion specific ISSMZ Sensor networks have a wide variety of applications, ranging from environmental data collection, seismic detection, health monitoring, battlefield surveillance and enemy tracking, etc. It is unlikely that there will be a general solution to satisfy all diversified or even conflicting requirements of all applications. For example, the tracking and surveillance applications need the detected events to be processed and reported within a very short period at high delivery rate. On the other hand, environmental data collection applications can tolerate a certain amount of data loss and larger transmission latency in order to maximize the network lifetime. Generally, sensor network protocols are optimized at all layers to meet the demands of the specific applications. Enemy Consinerations: Energy is a scarce resource for wireless sensor networks. Most sensor nodes are powered by non-rechargeable batteries, except for a few mobile sinks [2]. Therefore, it is essential to conserve computation and communication energy dissipation at all layers, from hardware design, medium access control, to network routing. Communication is usually the most energy extensive operation. The network stack should attempt to minimize energy usage, either by eliminating communication, or by turning off the radio when no communication is needed. The network lifetime has a strong dependency on the battery lifetime. In multiple hop communications, sensor nodes have a dual role: data generation and data forwarding. Node failure can break routing paths, and incur network topology re-organization. In some severe cases, it can even block all paths to data sinks, causing network partition. The need to efiiciently utilize energy has a deep impact on the system architecture and protocol design. Scalability: The number of nodes deployed in sensing fields may vary from hundreds to thousands, or even hundreds of thousands, depending on the type of applications. For example, the networks for environmental monitoring are likely to cover miles of target area comprising potentially thousands of sensor nodes. The huge number of devices not only provides increasing ability to instrument the physical world with great fidelity, but also poses a big challenge for network design. Frequent network-wide data dissemination or routing path establishment is not appropriate, which not only consumes the limited available bandwidth, but also incurs a heavy energy burden on networks. Scalability requires information exchange and node coordination to be local. Minimal information should be disseminated towards the whole network. Data DeliverLModels.' Depending on the application types, the data delivery model from data sources to sink nodes can be categorized as continuous, event-driven, query-driven, or hybrid [3]. In the continuous data delivery model, sensor data sources periodically send data to sink nodes. Typical applications include environmental monitoring and health monitoring. In event-driven and query-driven models, the transmission of data is triggered when interesting events are detected by sensor nodes or queries are generated by sink nodes to measure sensing attributes. This type of data delivery model is suitable for time-critical applications, such as intrusion detection and collaborative disaster management. Some networks apply a hybrid model of continuous, event-driven delivery model. Network architecture, especially routing protocols, is highly influenced by the data delivery model. It is suggested in [4] that for habitat monitoring applications, a hierarchical routing protocol is the most desirable to suppress the redundant data. In-network Da_ta Processing: Low cost and low energy supply require redundant deployment of wireless sensor nodes in many applications. This fact, together with simultaneous event detection by multiple nodes, results in significant redundant data being generated. Highly correlated data from multiple nodes in close proximity should be aggregated before being reported to sink nodes. Considering the fact that computation is much less energy consuming than communication [5], it is desirable to perform in-network data processing instead of sending all raw information to sink nodes. Data fi'om different sources can be combined based on aggregation functions, such as suppression, min, max, and average [6]. Signal processing techniques can also be utilized, which are capable of producing more accurate signals by beamforming or noise reduction [5]. In-network data processing has multiple impacts on routing protocols, such as improving energy efficiency and reducing traffic load. Network Dynamch: wireless sensor nodes are assumed to be stationary in most network architectures. However, there are some applications which require mobile sink nodes to improve performance [2]. Routing in the mobile context is more challenging as route stability becomes an important issue. The sensing event can be dynamic as well, based on the specific applications. For instance, in target detection/tracking applications, the movement of a target might incur dynamic event generation. Dynamic events generally require in-network data processing; otherwise extensive traffic will be route back to base stations. Even if all the nodes in the networks are static, yet the network topology can still be dynamic because of node failure or energy drainage. Frequent topology changes impose additional overheads on the network layer, as routing paths need to be rediscovered. @alitv of Service: In some applications, data should be delivered within a certain period of time from the moment it is sensed; otherwise the data will be useless. Therefore bounded latency for data delivery is another condition for time-constrained applications. However, in many applications, conservation of energy, which is directly related to network lifetime, is considered relatively more important than the quality of data sent. As the energy gets depleted, the network may be required to reduce the quality of the results in order to reduce the energy dissipation in the nodes and hence lengthen the total network lifetime. Hence, energy-aware routing protocols are required to capture this requirement. 1.2 Wireless Sensor Network Applications The majority of wireless sensor network applications can be categorized into three classes: environmental data collection, event-based monitoring and object tracking. 1.2.1 Environmental Data Collection Environmental data collection applications represent a class of sensor network applications with enormous potential benefits for scientific research communities and the society. It provides long term data collection at scales and resolutions which are very difficult to obtain in traditional methods. The intimate connection with the physical environment allows sensor networks to provide localized measurement and detailed information. Researchers collect sensor measurement from an environment over a long period of time in order to detect data trends and interdependencies. The measurement would be gathered from hundreds of points spread over several miles in field environment The data collection duration would last over several months or even years in order to get long-term and seasonal trends. For the data to be meaningful it may require to be measured at regular intervals and the nodes would remain at known locations. The data collection tasks should be flexible, and adjustable by user requests. For example, researchers should be able to send out sensing tasks, asking to collect information about temperature at area “A” for 2 weeks at an interval of 10 minutes. And later a new task to know humidity measurement at area “B” for half year at an interval of 20 minutes may also be generated. The sensor networks should be flexible to support various and dynamic data collection applications. At the network level, the environmental data collection application is characterized by having a large number of nodes continually sensing and transmitting data back to a set of base stations. These networks generally require very low data rates and extremely long lifetimes. Depending on application’s requirements for sensing coverage and resolution, the network size may vary from hundreds to hundreds of thousands of nodes. While the base stations for data collection and aggregation may be powered by wire connections, the majority of nodes in the networks are powered by batteries, which are not replaceable. Environment data collection applications usually use tree-based routing strategy. Data is transmitted from child nodes to parent nodes along the tree, till reaches the rooted base stations. Although the leaf nodes may generate a small amount of data, the nodes with lots of descendants need to receive and relay considerably large amount of data. Therefore, these nodes are more likely to be energy bottlenecks. In order to conserve communication energy, the radios of nodes should be turned off whenever there is no ongoing traflic. In this case, an energy-efficient TDMA type Medium Access Control is desired. For sensor networks, it is expected that some nodes will fail over time, due to hardware failure or energy drainage. And it is also possible for new nodes to be added into the network to replace failed ones or to extend coverage. In both scenarios, the networks need to reconfigure to discover the topology and reestablish routing paths. The network reconfiguration should be local and infrequent to avoid high energy dissipation due these overhead. One of the typical environment data collection applications is the Great Duck Island (GDI) system. Researches deployed a mote-based tiered sensor network on Great Duck Island, Maine to monitor the behavior of storm petrel from collected information on light, temperature, infrared, humidity and barometric pressure [7]. 32 motes are deployed at area of interests, which are grouped into sensor patches. These sensor patches send data to gateways through multiple hops, and then the gateways relay the data to a base station through an 802.11 wireless network. At the base station, the data is updated to database in Berkeley every 15 minutes through satellite links. Users can access the data either from the remote server in Berkeley or read local data at Great Duck Island by a Compaq iPaq PDA. In general, environmental data collections applications are characterized by long network lifetime, low data rate, relative static network topology, and tolerable delay. 1.2.2 Event-Based Monitoring Another class of sensor network applications is event based monitoring. Compared with environmental data collection applications, event detection monitoring applications are not intended to collect data, instead they aim to detect whether predefined events happen in the environment, and report these events to users. The majority of energy dissipation here is not for communication, but for event detection and in-network processing. The real-time and reliable communication of detected events is the most important requirement, as the events are expected to happen very infrequently and the consequences of event loss could be severe. All these requirements have a significant impact on network architecture design. When sensor nodes detect an event, a message needs to be generated and reliably forwarded to the base station in real time. The allowable message latency is determined by the nature of different applications. For example, the latency for a detected fire alarm in forest can be within minutes, while enemy intrusion detection must be reported in seconds. In the latter case, the network should be able to respond quickly to detected events and forward data. The traflic pattern of event detection monitoring applications is “bursty”. Most of the time events do not happen in the networks, and sensor nodes only need to periodically check neighbor information to ensure nodes’ ftmctionality. However, once an event is detected, a significant amount of traffic will be generated in a short period. Nodes need to collaborate with each other track the changes, and report messages to the base stations. It is very likely that multiple nodes will detect the same event simultaneously; therefore, in-network data processing may be employed to enhance message accuracy and reduce traffic load. It is also important to confirm the functionality and aliveness of sensor nodes. If nodes had suffered failure and could not monitor sensing areas, it would represent a security violation that need to be reported. The network should have the ability to ensure the monitoring coverage of sensor nodes; otherwise it can not fimction properly. This requires coordination between neighbor nodes. The topology control of an event detection monitoring network will look quite different from that of a data collection network. A large portion of the energy will be spent on confirming the firnctionality of neighbor nodes and performing event detection. The data message transmission will consume a small fraction of the network energy. Firebug [8] project is a wildfire monitoring application, based on Berkeley motes. It is composed of environmental sensors collecting temperature, relative humidity and barometric pressure with an on-board GPS unit attached to a wireless, networked mote. The motes communicate with a base station, which sends the collected data to a database server. The data can be accessed using a browser-based web application or other application capable of communicating with the database server. Performance of the monitoring system during two prescribed fires at Pinole Point Regional Park (Contra Costa County, California) is promising. Sensors within the fire zone recorded the passage of the flame front before being scorched, with temperature increasing, and barometric pressure and humidity decreasing as the flame front advanced. 1.2.3 Object Tracking A third class of sensor network applications is to track an object status in real time. The most visible and publicly discussed example is the usage of RF ID (radio frequency identification) tags in supply chain management. US Department of Defense demands the support of RFID technology from their supplies within the next few years. Delta Airlines with US Transportation Security Administration successfully performed a pilot project for baggage handling through RFID at the airport of Jacksonville, Florida [9]. And Wal-Mart Stores Inc. mandated that its top 100 US supplies tag all cases and pallets carrying merchandise by January 2005 [10]. In the traditional inventory system, when merchandise passes a checkpoint, its barcode is scanned by a reader, so that the system can track the data flow through different checkpoints. However, it is time consuming and inefi'rcient to track the objects 10 because of the inflexible barcode scanning processing. And it is impossible to provide real-time status for the merchandise. With wireless sensor networks, objects can be identified and tracked by attaching RFID tags. The tags can be active, which are able to announce the presence of the devices. Through wireless communication with RFID readers, the objects can be easily and quickly tracked without any physical contact. A database can record the location of tracked objects and check whether they are damaged and need replacement. With the system, it is possible to query and process the current statuses of objects at very low cost. Then the business applications respond to these inputs and arrange the corresponding actions, such as replacing component or requesting the delivery of additional products. To be cost effective, the device is a very low power, low performance sensor unit [11]. It includes physical sensors, a low-power micro controller, limited memory, and a low-power, short-range wireless radio. The duty cycle is expected to be 0.01%, which means the devices are in sleep mode about 99.9% of the time. The medium access control protocols should be very energy efficient, and adaptive to topology changes. In addition, the tracked objects may continually change positions, it is important for the network to efficiently detect the presence and leave of the objects. 1.2.4 Other Applications In general, applications may require multiple features. For example, in an environmental data collection application, researchers may generate requests to get periodical monitoring data for wild animals. In the meantime, they are interested to detect the environment anomaly, such as flood, fire, etc. Once an anomaly happens, the 11 detected events will be reported back to the base stations quickly and reliably. It is also possible to track the movement of specific animals like giraffes, in order to understand their behavior in the natural habitats. The networks should be flexible enough to support all the requirements in a single system. There are other types of applications whose major challenges are on micro-electro-mechanical system (MEMS). In health application [12], researchers are working on artificial retina. Retinal prosthesis chip consisting of 100 micro-sensors are built and implanted within human eye. It allows patients with no or limited vision to see at an acceptable level. The wireless communication is needed for feedback control, image identification and validation. The combination of sensing, micro controller and wireless communication provides tremendous potential applications for wireless sensor networks. These networks perceive the physical environment in ways that were not possible previously. They not only offer the potential to advance many scientific pursuits, but also provide a platform for enhancing large forms of productivity in the industry of manufacturing, agriculture, construction and transportation. 1.3 Research Issues on Energy Efficiency and Network Scalability Although wireless sensor networks have already demonstrated a promising future, there are a number of challenges to be overcome before they become reality in a wide range of applications. The issues include fault tolerance, network scalability, cost, hardware, topology management, environment changes and power consumption [13]. In this work, we mainly consider the challenges of power consumption and network 12 scalability from the medium access control and the networking layers. 1.3.1 Energy Efficient Medium Access Control Energy is the most constrained resource for wireless sensor networks. From the development history of battery, it is not expected to see significant breakthrough in next few years. Therefore, it is important to consider the energy efficiency from all communication layers. This constraint and the associated computation and storage limitations lead to a new set of communication architectural design. Many wireless devices, such as cell phones and PDA, can reduce their power consumption through the use specialized communication hardware in ASICs that provide low-power implementations of communication protocols [14]. However, the strength of wireless sensor networks is their flexibility and universality. The wide range and different nature of applications makes it difficult to develop single protocol stack to suit for all requirements. Therefore ASIC hardware design can not be applied for all applications. A wireless sensor network platform must provide support for a suite of application-specific protocols that drastically reduce node size, cost, and power consumption for their target applications. In wireless sensor networks, communication is significantly more expensive than computation. Therefore, it is very important to reduce energy dissipation in communication. In a shared medium MAC, the non-essential energy expenditures are contributed [16] by the following four sources. (1) protocol overhead: to execute resource control, a MAC protocol has to incur some overhead, besides the basic data; (2) collisions: retransmissions due to collisions lead to unnecessary energy consumption; 13 (3) overheating: nodes expend energy to receive packets for which they are not the intended receivers; (4) idle listening: in protocols that do not have any centralized transmission slot allocations, nodes have to always keep their wireless radio interface powered on for all possible receptions. This consumes power even when a node does not transmit or receive for a long time. Research has shown [16, 17] that idle listening accounts for most of the energy consumption at low traffic situations, which are prevalent for lots of wireless sensor network applications. Fortunately, many wireless sensor network applications are delay tolerant and require low bandwidth. It is possible to trade classical performance parameters (throughput, latency, fairness) for better energy efficiency. At MAC layer, sensor nodes should be able to reduce the radio duty cycle to improve the lifetime of networks. In this work, we propose the Self Reorganizmg Slot Allocation (SRSA), for implementing spatial TDMA type MAC protocol for sensor networks with clustering abstraction. The details of our design are given in chapter 2. 1.3.2 Scalable Routing for Very Large Sensor Network Typically, the number of sensor nodes deployed in fields may be on the order of hundreds or thousands. Depending on the applications, it is possible for the number to reach even millions. These very large sensor networks not only provide a revolutionary way to perceive physical world with great fidelity, but also pose a big challenge for network protocol design. Routing protocols for Ad hoc networks generally use either distance-vector or link-state routing algorithm [18]. Both types of protocols find shortest paths to 14 destination nodes. In distance-vector routings, a vector containing the cost (e.g., hop distance) and the path (next hop) to the destination is kept and exchanged at all the nodes in the network, when routing discovery is generated. The link-state routing algorithms periodically flood link information about its neighbors to maintain the global network topology at each node. In large networks, the transmissions caused by routing overhead can be overwhelming. These messages will consume a large part of the bandwidth and consequently block applications. They can have a series of adverse effects including frequent packet drops and increased latency. A more detrimental effect is the increased energy overhead on the resource constrained sensor nodes. Thus, reducing routing control overhead becomes a key issue in achieving the routing scalability. Hierarchical routing protocols relieve the scalability problem by organizing nodes in groups and assigning nodes different ftmctionalities inside and outside of a group. Both routing table size and the number of packets exchanged are reduced as only parts of the nodes in the network are involved in the route discovery process. An example of hierarchical routing is the Internet hierarchy, which has been practiced in wired networks for a long time. In wireless hierarchical routing, the most popular way of building hierarchy is to group nodes geographically close into clusters. Each cluster has a leading node (cluster head) to communicate to other nodes on behalf of the cluster. The cluster heads and the corresponding nodes connecting them build a “backbone” network. Routing discovery messages are exchanged on the “backbone” network. However, in the wireless network, as there are no strict hierarchy management like 15 Internet, the efficiency of hierarchical routing is very limited. To discovery paths, the routing messages need to flood the whole “backbone” network, which includes 20%~40% of the total network nodes [19]. Hierarchical routing protocols suffer from the scalability problem in large networks. Geographic routing protocols take a different approach to address the scalability problem. Sensor nodes are assumed to know their locations, either from Global Position System (GPS) devices or location identification algorithms [42]. With the knowledge of node locations, routing can be more effective and scalable at the cost of overhead incurred by exchanging neighbor coordinates. The main idea of geographic routing works as follows: when a node receives a packet with destination’s location, it chooses one of its neighbors whose is geographically closest to the destination, and forwards the packet to it. The process repeats at each intermediate node till the packet reaches the destination. However, if the packet arrives at a dead end (all the neighbors are farther way from destination than itself), a detour search is incurred to find the path to destination. In regular topologies where detour search can be avoided, geographic routing perhaps is one of the most efficient protocols, as it only needs to exchange l-hop neighbor coordinates information, which does not require network-wide route discovery. However, in sensor networks it is not financially feasible to equip every cost-constrained sensor node with GPS devices. And in the events of frequent detour searches, the additional control overhead lessens the advantages of these protocols. Frequent detour searches are usually caused by irregular terrains and other administrative and security-based network partitioning. 16 To solve the routing scalability problem, we propose a routing solution that achieves control plane scalability by avoiding network-wide route searches without relying on geographical information. The key idea is to shift part of the route computation to an “off-networ ” routing server. This server is assumed to be non energy-constrained, and it is used for performing “coarse grain” global route computation. The “fine grain” local routing is performed locally within the sensor nodes in a distributed manner. This hybrid server-based and distributed approach can significantly decrease the requirements for routing control message exchange. 17 Chapter 2 A Self Reorganizing Slot Allocation Protocol for Multi-cluster Sensor Networks 2.1 Introduction and Motivation Sensor networks typically consist of very large number of low cost sensor nodes which can collaborate among each other to enable a wide range of applications such as environmental and health monitoring, security and intrusion detection, and collaborative disaster management. Communication protocols for traditional networks are divided into several layers, and these protocols are designed for flow-oriented transactions in which data is generally routed from one or multiple predefined sources to one or multiple predefined destinations. Sensor networks, however, are data-centric in which a piece of data is generally routed based on its content and context rather than on predefined source-destination flows. In addition, communication in sensor networks is highly influenced by their various constraints, including limited energy supply. Therefore, it is crucial for the sensor network protocols to be energy efficient in order to extend the network lifetime. Traditional wireless MAC protocols such as IEEE 802.11 [2 1], CSMA and HiperLan-II [22] are designed for optimizing throughput, latency, and fairness without specifically concentrating on their energy usage. In sensor networks, Since energy efficiency is one of the primary requirements, these protocols are not directly useable without substantial modifications. Therefore, new energy-aware MAC pr Otocol models are necessary for sensor networks. Also, while the traditional “split” layered protocols have advantages in terms of funCational abstractions and interoperability; they often pose severe encumbrances for 18 energy optimization due to their functional generality and implementation overhead. To address this, cross-layer designs are required so that instead of being functionally independent, the protocols at different layers are designed to be interdependent and can be optimized together to achieve better energy performance. As we discussed in chapter 1.3, energy is the most constrained resource for wireless sensor networks, and communication is significantly more expensive than computation. It is essential to reduce to the energy dissipation in wireless communication. In this chapter, we present an energy efficient MAC protocol, Self Reorganizing Slot Allocation (SRSA), for implementing a spatial TDMA protocol in sensor networks. While reducing energy consumption in MAC layer, SRSA also leaves room for cross-layer performance optimization in routing layer. The central idea behind cross-layer optimization is to adapt routing based on cluster-based slot allocation by the SRSA MAC. Here we will only concentrate on the MAC layer aspects of SRSA. In a shared medium MAC, the non-essential energy expenditures are contributed [16] by the following four sources. (1) protocol overhead: to execute resource control, a MAC protocol has to incur some overhead, besides the basic data; (2) collisions: retransmissions due to collisions lead to unnecessary energy consumption; (3) overhearing: nodes expend energy to receive packets for which they are not the intended receivers; (4) idle listening: in protocols that do not have any centralized transmission slot allocations, nodes have to always keep their wireless interface powered on for possible receptions. This consumes power even when a node does not transmit or receive for a long time. Research has shown [16, 17] that idle listening l9 accounts for most of the energy consumption at low traffic situations, which are prevalent for lots of wireless sensor network applications. A significant amount of work has been done [16, 17, 21-30] on decreasing idle listening. The main idea behind this research is to power off network interfaces when possible. In effect, a node’s operating time is divided into active-sleep cycles. Each cycle consists of an active duration for regular transmissions and receptions and a sleep duration during which the network interface is powered off for saving idle listening consumption. The ratio of active period to active-sleep cycle duration is referred to as the duty cycle of operation. Smaller the duty cycle, lower the consumption due to idle listening. The main goal of all protocols in this category is to operate at the smallest possible duty cycle while being able to support application loading requirements. TDMA MAC protocols also belong to this category because of their built-in active-sleep duty cycles that can be leveraged for limiting idle listening. Although this makes TDMA a natural choice for sensor MAC, successful implementation would require some form of spatial TDMA [31] in the presence of overlapping MAC clustering. In such deployments, the main challenge is to devise TDMA slot allocation mechanisms with the goal of minimizing allocation collisions across overlapping clusters. A straightforward solution [5] to this problem is to use Code Division Multiple Access (CDMA) or Frequency Division Multiple Access (FDMA) across clusters and then to run independent TDMA schedules within each cluster. However, using CDMA or FDMA incurs costs in terms of spectrum usage as well as network interface complexity for low-cost sensor nodes. 20 The design objective of our work is to devise mechanisms for minimizing inter-cluster TDMA interference without actually using CDMA or FDMA. To accomplish this, one approach will be to use a centralized scheduler that can perform TDMA slot allocations for the entire network based on cluster boundaries and overlapping across neighbor clusters. To avoid the scalability concerns of such centralized mechanisms, we propose a distributed Self Reorganizing Slot Allocation (SRSA) protocol that operates at the cluster level and does not rely on any information beyond a cluster. The main idea is to initiate communication with a reasonable initial TDMA allocation and then, adaptively change the slot allocation schedule locally based on feedback derived from collisions experienced by the local nodes. SRSA slot allocation is executed adaptively relying only on local information. The main contribution of this work is to demonstrate that for low sensor loading scenarios, the SRSA protocol is able to minimize TDMA slot interference across neighboring clusters without relying on CDMA, FDMA or any global synchronization mechanisms. In Section 2.2, we describe the related work on energy efficient MAC protocols. Then in Section 2.3, we elaborate the design of the proposed SRSA MAC protocol. In Section 2.4, we describe our simulation setup, followed by a detailed analysis of the experimental results. And we summarize our work in Section 2.5. 2.2 Related Work Active-sleep duty cycle based mechanisms for limiting idle listening have been applied to both contention-based as well as contention-free MAC protocols [16, 31-36]. In contention-flee protocols like TDMA, since transmission and reception time slots 21 are pre-allocated, and nodes can sleep during the known inactive periods, they are inherently more energy efficient compared to the contention-based protocols. Ye et a1. propose S-MAC [16], in which the basic idea is to synchronize nodes in an active-sleep cycle so that the nodes can communicate during the active period and turn their interfaces off during the sleep period. During the active periods node execute an 802.11-like light weight CSMA-CA MAC protocol. Sensor nodes form virtual clusters so that all nodes within a virtual cluster maintain the same active-sleep schedule. While the active duration determines allowable traffic load, the sleep duration determines the packet delivery latency. By decreasing the active-sleep duty cycle, this protocol can trade energy for latency and channel bandwidth. SMAC suffers from the following shortcomings. First, since the basic medium access mechanism is contention-based, due to channel contentions the protocol cannot guarantee bounded delivery latency at high loads. Second, nodes at boundaries of the virtual clusters require maintaining multiple schedules, one for each cluster. Maintaining multiple schedules requires a node to remain awake during the active periods for all its maintained schedules. As a result, nodes with multiple schedules can have significantly larger energy drainage compared to the nodes with only one schedule. An alternative way to introduce periodic active-sleep cycles is by using the concept of wake-up radio [17]. In this protocol, a node normally sleeps, but when it has data to transmit, it first sends a signal through another channel to wake up the receiver node. Since the only function of the radio for wake-up channel is to wake up sleeping nodes (no data processing), the hardware for that interface can be simpler and less energy 22 hungry. Nevertheless, this design adds an extra transceiver which is an added cost and complexity burden for low-cost sensor nodes. CSMA with preamble sampling [32, 33] incorporates the concept of active-sleep cycles from the receivers’ perspective. The idea is to precede each transmission with a long preamble of duration Tp. A node wakes up every Tp duration to sense the channel. If no preamble is detected within Ts duration, the node goes back to sleep. If a preamble is detected, then a node continues to listen until a packet has been received or a timeout occurs. This allows receivers to sleep most of the time when the traffic is low. When the channel is idle, the approximate active-sleep duty cycle of the radio is Ts / (Ts + Tp). A notable feature of this protocol is that unlike in SMAC, the duty cycle of operation does not have to be pre-defined based on anticipated traffic load. It is automatically adjusted based on the instantaneous loading conditions. Preamble sampling shifts idle listening energy burden from receivers to senders in the form of energy expenditure for transmitting long preambles. The situation worsens with increased collisions and retransmissions at higher loads. Additionally, like other contention-based protocols, it suffers from the unbounded delivery latency problem. Under low traffic loads, all these protocols work well. But when load increases, these contention—based protocols degrade drastically because of collisions and subsequent retransmissions. For higher load situations, TDMA based protocols can offer better channel utilization [34] and energy performance due to their inherent ability to follow active-sleep cycles based on periodic slot allocations. The main challenge for using spatial TDMA in a large sensor network is to come up with slot allocation schedules for 23 maximum channel utilization with minimum packet delivery latency. Most spatial TDMA protocols do not scale well, since they assume the knowledge of the global topology. Scalability can be achieved by partitioning a network into smaller clusters and running independent TDMA framing within each cluster. To avoid allocation collisions across neighbor clusters, the use of CDMA (or FDMA) has been suggested in [5]. However, using CDMA or FDMA incurs costs in terms of spectrum usage as well as network interface complexity for low-cost sensor nodes. The objective of our work is to devise a mechanism for minimizing inter-cluster TDMA collisions without using CDMA or FDMA. 2.3 Self-Reorganizing Slot Allocation Protocol (SRSA) Design of SRSA is targeted to meet the following two requirements. First, the slot allocation process should be online and adaptive, with detected collisions as a feedback parameter. Second, the protocol has to operate locally without having to rely on any inter-cluster communication, synchronization and global timing. 2.3.1 Network Model and Assumptions In our network model (see Fig 2.1) data is sensed from a sensor field periodically and delivered to monitoring applications through base stations and a wired network. It is assumed that each sensor node can communicate directly with at least one base station. For scalability, however, the network is self-organized into multiple clusters [5], and within each cluster a sensor node is autonomously elected as the cluster head of that cluster. A cluster head acts like the gateway between the nodes within its cluster and the base stations. 24 From an energy standpoint, the advantages of clustering are as follows. First, by routing all data through the local cluster heads, the nodes avoid high-power long distance wireless transmissions to the base stations. Only the cluster heads have to transmit under such distances. A cluster head can fiuther reduce the transmission energy expenditure by aggregating the collected data from its cluster before relaying them to the base stations. This reduces the overall network-wide transmission energy expenditure. Since the monitoring applications are often interested only in geographically aggregated rather than per-node data, aggregation at cluster heads is highly desirable for extending the lifetime of sensor networks. Since the cluster-heads spend more communication energy than the regular sensor nodes, the responsibility of cluster-head has to be rotated [5] across all the sensor nodes for energy load-distribution. Ap " 88-2 pnwfion I J .- JR;- I A \- s 2' l'r ‘ “CH" fir: " ‘ - '... ..-“-'..' "30‘“. ‘1 F 3 ‘ _ . .,'-..‘ can-1‘2?! w n ‘ ('7‘ / ....'. .. ”A ,\L ..»..n ...- - ‘k. -, 335° ‘1 \r Station-1 "s, "x. 'o ‘u.. ’s Cluster Heads Fig. 2.1: Target sensor network and data flow model Successful operation of this model would depend on three crucial protocol components, namely, self-reorganizing cluster formation, autonomous cluster-head 25 election and energy-efficient medium access control. This work proposes a new MAC protocol while relying on existing cluster formation protocols [2, 25, 26, 27, 29] and cluster head election mechanisms [2] available in the literatures. Note that we provide a MAC solution only for the intra-cluster communication. A separate mechanism will be necessary for the base station to cluster head communication. 2.3.2 SRSA Protocol Overview After the clusters are self-organized and cluster heads are elected, each cluster maintains a local TDMA MAC flame. Each regular sensor node is allocated TDMA slots for uplink transmissions as well as for downlink receptions. After a node receives its slot allocation, it remains active during the allocated slots and sleeps during all other slots. In SRSA, we assume that the packets are of fixed size and the TDMA flame durations in all clusters are the same and it is pre-configured. However, the frame timing across the clusters is assumed to be asynchronous and the number of nodes in different clusters is not assmned to be equal. Nodes within a cluster are time synchronized through their cluster-head. TDMA slot duration is equal to the packet time plus a short duration for carrier sensing before transmission. Data is exchanged only between a sensor node and its local cluster head. The baseline SRSA protocol is as follows. Step-l (Initialization): A cluster head initially makes a random TDMA slot allocation to its cluster nodes. Stein-2 (Carrier sensing): Before transmitting, a node first senses the carrier to see if 26 there is an ongoing transmission due to an overlapping allocation in any of the neighbor clusters. If the channel is found flee then the packet is transmitted. Otherwise, a Carrier Sense Collision (CS-Collision) is declared and the packet is buflered until the next allocated slot for that node. A CS-collision indicates that there exists at least one overlapping allocation in one of the node’s neighbor clusters. An example CS-collision situation will arise in the network in Fig. 2.2 when node C in cluster-1 and node D in cluster-2 are allocated overlapping TDMA slots. Since C and D are within each otherS’ transmission range, the node which starts transmission later will detect a CS-Collision. SteL3 (CS-Collision Feedback): Upon detection of a CS-collision, a node sends that information to its cluster head through the next successful uplink packet transmission. A bit in the packet header is designated for carrying this information. This method of learning about a CS-collision by a cluster head is referred to as active detection. Another way to infer about a CS-Collision is by passive detection. If a cluster head observes that although uplink slots have been allocated to it, a node in the cluster is not transmitting uplink packets for a large number of flames, then a passive detection of CS-Collision takes place. Passive detection works better in higher loading situations. Step-4 (HIV-Collision Detection): Consider a situation in Fig. 2.2 when node C in cluster-1 and node E in cluster-2 are allocated overlapping TDMA slots. Since these two nodes are outside each others’ transmission range, instead of CS-Collision, a Hidden Collision will occur at the head of cluster-2. Unlike CS-Collisions, the HN-Collisions cause packet drops and can be directly detected by the affected cluster heads. It cannot be detected by the colliding nodes. 27 Step-5 (Slot Allocation Reorganization): At the end of each TDMA frame, a cluster head examines all TDMA slots for CS-Collisions or PIN-collisions. Then the cluster head executes a reorganization of the local slot allocation so that the number of CS and HN collisions can be reduced in subsequent TDMA flames, and eventually minimized to a stable range. The reorganized TDMA schedule is then downloaded by the cluster head to its cluster members at the beginning of the next frame. A possible schedule download mechanism for a cluster head would be to broadcast a bitmap specifying the new slot allocation schedule. Depending on the size of the bitmap, it can be downloaded either as a part of the header of a downlink user packet or as a separate downlink control packet Since all clusters work in a distributed way, missing a schedule only increases the convergence time for SRSA to reach a stable state. Steps 3, 4 and 5 are iteratively executed to maintain stable and optimal allocation. Details of the allocation reorganization process in Step 5 and its impacts are described in the next few subsections. Cluster-1 Cluster-2 '——-- ~~~~~ I’ ‘ Physical Connectivity Fig. 2.2: TDMA Collisions across overlapping clusters 2.3.3 Inter-cluster Allocation Dependency All possible overlapping slot allocations for the network in Fig. 2.2 are summarized in Table 1. To clarify, for example if node C and node B are allocated overlapping slots 28 then an HN-Collision will occur at the head of cluster-2. Although the goal of an optimal allocation is to maximally utilize the collision flee combinations, depending on the relative locations of the nodes it may or may not be possible to completely avoid collisions. For example, with the random allocation in Fig. 2: a, only nodes A and C are able to successfully transmit their uplink packets. Uplink transmissions from all other nodes face either CS or HN collisions. Note that the CS-Collision for the {B, D} combination affects only node B because of the relative timing of the flames of the clusters. With the shown timing, node D’s slot always starts before B’s, and as a result B detects a CS-Collision during the carrier sense at the beginning of its allocated slot. :l 3 I0 V Cluster-1 CABCABCAB Cluster-2 E F D E F D E F D a) Random Allocation: HN-Collision due to {C, E} and {C.F} at the head of Cluster-2; CS-Collision due to {8.0} at B Cluster-1 CBACBACBA Cluster-2 DFEDFEDFE l b) Optimal Allocation: HN-Collision due to {C.F} at Cluster-2 head; CS-Collision due to { C. D} at C Fig. 2.3: Example slot allocations for network in Fig. 2.2 In an optimal allocation (see Fig. 2.3 :b), the CS-Collision due to {B, D} combination in Fig. 2.3:a has been removed. Also, the HN-Collision due to {C, E} combination in 29 Figure 2:a is removed at the expense of a new CS-Collision due to the allocation combination {C,D}. Since HN-Collisions result in packet drops and CS-Collisions result in increased delivery latency due to transmission deferments, it is possible to trade better drop rates with larger delivery latency in a SRSA allocation paradigm. Overall, the new allocation arrangement results in one less HN-Collision compared to that in Fig. 2.3: a. This turns out to be one of the best allocations that can be achieved for the network in question. The collisions here cannot be completely removed because of the fact that C’s transmissions can be always be heard by the heads of both the clusters. Allocation Overlapping Slots to Node-pairs Type in Neighbor Clusters Collision Free {A,D}, {A,E}, {A,F}, {B,E} , {B,F} CS-Collision {B,D}, {C,D} HN-Collision {C,E} at E’s cluster head {C,F} at F’s cluster head Table 2.1: Allocation possibilities for the network in Fig. 2.2. 2.3.4 Frame Scaling for Collision Reduction To reduce collisions in networks with nodes at cluster boundaries (e.g. the network in Fig. 2.2), we introduce the concept of flame scaling, according to which the TDMA flame duration is scaled up so that there are more slots in a flame than the traffic in that cluster. The ratio of the number of available slots in a flame and the amount of traffic (slots per flame) in a cluster is referred to as the Scaling Factor (SF). The motivation for flame scaling is to introduce unused flee slots which can be used for alleviating both CS and HN collisions during the allocation reorganization process. With scaling factor 2, a collision free slot allocation for the network on Fig. 2.2 is shown in Fig. 2.4. 30 Frame scaling has the following shortcomings. First, by inserting free slots, the effective channel capacity of a cluster is reduced. Second, with increased flame duration, the average packet delivery latency goes up. Conceptually, flame scaling serves the same purpose as that of active-sleep cycles in SMAC [16] and preamble sampling in [32, 33]. The flee slots in sealed flames extend the sleep duration of a node at the expense of lowering the allowable loading range. In effect, flame scaling decreases TDMA’s inherent duty cycle by the flame scaling factor SF. Note that in our model, SF can be different in different clusters depending on the nodes within that cluster. The flame lengths for all clusters are assumed to be the same. Time _ Frame (SF = 2) Cluster-1 A B C Cluster-2 E F D Collision-free \ allocation (SF = 2) Free Slots Fig. 2.4: Use of flame scaling for collision reduction 2.3.5 Allocation Reorganization Algorithm For a given cluster, its future allocation schedule depends on its cmrent schedule, its neighbors’ current schedules, scaling factor and the CS/HN collision rates. The following mechanism is used by a cluster head for allocation reorganization mentioned in Step 5 in Section 2.3.2. gar-J: At the end of a flame, a cluster head examines .all the slots and mark their status as one of F (Free), U (Uncontested: there was no collision detected for this slot), 31 C (CS-Collision detected) or H (HN-Collision detected). M: For each H slot, the cluster head looks for an F slot and, if found, the H and F slots are swapped between the previously allocated nodes. M3 If for an H slot, an F slot is not found then the cluster head looks for a C slot and, if found, the H and C slots are swapped between the previously allocated nodes. W: If for an H slot, an F or a C slot is not found then the cluster head looks for a U slot and, if found, the H and U slots are swapped between the previously allocated nodes. M1 For each un-swapped C slot, the cluster head looks for an F slot and, if found, the C and F slots are swapped between the previously allocated nodes. S_teL6: If for a C slot, an F slot is not found then the cluster head looks for a U slot and, if found, the C and U slots are swapped between the previously allocated nodes. SRSA initialization and end-of-flame execution logic is summarized in the Pseudo code 1. Iterative and distributed execution of this protocol logic at the cluster heads leads to a stable slot allocation with minimized HN and CS-Collisions. As mentioned earlier, one of the main objectives of SRSA is to reduce HN-Collisions (for reducing packet drops) more aggressively than the CS-Collisions, which are responsible for packet delivery delay. This has been accomplished by first swapping the ‘H’ slots with ‘F’, ‘C’ and ‘U’ slots in that order, and then swapping the remaining ‘C’ slots with ‘F’ and ‘U’ slots. This way, removal of the I-lN-Collisions is given a higher priority. Note that after exhausting the available ‘F’ slots, SRSA swaps the ‘H’ slots with available ‘C’ slots. The rationale behind this step is to flee a slot from HN-Collision by 32 exposing it to CS-Collision. The anticipated result is to reduce the overall packet dr0p rate (by reducing PIN-Collisions) at an expense of increased packet delivery latency (as a result of increased CS-Collisions). However, for the ‘C’ slots, the reverse (swapping it with ‘H’ slots) is not carried out; this is because SRSA does not ever attempt to reduce CS-Collisions at the expense of increased HN-Collisions. Finally, both HN and CS collisions are attempted to be removed by swapping ‘H’ and ‘C’ slots with the ‘U’ slots. Intuitively this seems to be a disruptive process of perturbing the already stabilized collision flee nodes. But depending on the topology, often times it turns out that while removing the HN or CS-Collision, such a swap does not expose the collision-flee node to a new collision. Even when it does, with sufl’icient number of flee slots (‘F ’ slots) in the flame, as the protocol converges, eventually the node becomes collision flee. Convergence characteristics of the protocol in such situations are presented in Section 2.4. 33 /* SRSA protocol accented at each cluster head */ Initialize cluster schedule based on random allocation; At the end of each SRSA/rame perform the following: { Process the CS-collision feedback received from the cluster nodes during this frame; Identijy all slots in the frame with hidden collisions; Mark slots as Free (F), Uncontested (ID, CS-Collision (C) and HIV-Collision (H) For (all H slots) { // try removing the fflV-Collisions first If (an F slot is available) { swap H and F slots; continue; } If (a C slot is available) { swap H and C slots; continue; } If (a U slot is available) { swap H and U slots; continue; } // done with all IflV-Collisions; CS-Collisions next For (all arr-swapped C slots) { If (an F slot is available) { swap C and F slots; continue; } If (a U slot is available){ swap C and U slots; continue; } } // done with CS-Collisions } // schedule for the next frame is ready Fig. 2.5: Pseudo code of SRSA protocol logic at each cluster head 34 2.3.6 Dimensioning Frame Scaling Factor While being able to reduce collisions, the tradeoffs for higher SF are increased latency and lower effective channel capacity. Therefore, it is necessary to dimension the value of SF such that it is no larger than what is needed to guarantee zero collisions at steady state. For a static network with given topology, its critical scaling factor SF critical is defined as the minimum SF at which there exists a zero-collision allocation combination across all the clusters. It turns out that finding the value of SF critical in an arbitrary network is NP hard. Therefore, in this work we compute the bounds for this quantity. For a given cluster i, we define the following parameters. local: number of nodes that belong to cluster i N :emote : Number of nodes that belong to other neighboring clusters and are located within the overlapping region of cluster head i N hflected : Number of nodes that belong to other neighboring clusters and are located within the transmission range of all nodes in cluster i. The concept and the defined quantities are illustrated in an example two-cluster network in Fig. 2.5. Note that these definitions are general and hold for larger number of overlapping clusters. Depending on the specific topology, in different clusters the quantities low, , Niemm , and dflected may be different. In the presence of cluster—overlapping, the minimum number of slots needed (for . . , i i zero collrsrons) for cluster 1 cannot be less than (N 10ml+ N remm ). Therefore, the 35 lower—bound of the critical scaling factor for cluster i can be computed as: heal (I) (N 10ch + N remote +1)/ local Note that the additional slot in the numerator is necessary to accommodate the inter-cluster flame asynchrony, as illustrated in Figs. 2.3 and Fig 2.4. Similarly, the minimum required slots for zero collisions cannot exceed (N [Zeal + N firmed). For a given topology, this corresponds to the upper-bound of the critical scaling factor, which can be computed as: SFU‘tical (i): (Nlocal + Nafl'ected + 1)/1vlo cal These bounds provide a guideline for network-wide scaling factor dimensioning in a multi-cluster network running SRSA. Consider a network with M clusters. Assuming equal flame size across clusters, the bounds of SFcritical across the entire network are: (Network): max(SFLB critical (1)) i = 1,2,3...,M critical crltica [=(NetW0rk) maX(SF critical(i)) i = I,2,3...,M Affected (shaded) area by \ Cluster—1 ' Cluster-2 Cluster-1 ForCluster-1: Nlocal= 6, Nremote_ — 1 and Naffected= 4 Fig. 2.6: Parameters for Frame Scaling Factor dimensioning 36 LB UB LB For the network in Fig- 2-6: SF critical (1) ’ SFhritical (1) 9 SPLritical(2) and U3 SFcritical(2) are manually computed to be 1.33, 1.83, 1.33 and 1.66 respectively. . LB UB Therefore, the network-wrde values of SEW“, and SF'Icn'fic-a] are 1.33 and 1.83, which are the maximum over both the clusters. For this network we have found the required actual SFcritical to be 1.33 (same as the value of 5175;“, ). This was experimentally found flom ideal allocation vectors which are established through a brute force search across the entire allocation space for the network in Fig. 2.6. 0.1 p 0.025 0.09 . .. . . :Ideal 0.08 o 0.02 i . . 8 .. . £0.07 a E 0.06 .5 0.015 £0.05 g {30.04 8 0.01 80.03 g 0.02 0.005 0.01 ' 0 0 0.9 1.1 1.3 1.5 1.7 1.9 0,9 1,1 1,3 1,5 1,7 1,9 Scaling Factor Scaling Factor 3) CS-Collisions b) Packet drop rate Fig. 2.7: Comparison between SRSA and ideal allocation For situations when SF > SF critical, SRSA will converge to a stable state. The convergence speed depends on network topology, SF, and the flafiic load. Collision comparison between SRSA and the manually found ideal allocation for the network in Fig. 2.6 is depicted in Fig. 2.7. As expected, after the scaling factor reaches SFcritical (1.33), both CS and HN collisions are completely removed by the ideal allocation. Note that the ideal allocation trades latency for lower packet drops. That is why its CS-Collision (which is responsible for latency) is higher in Fig. 2.7 (a). To summarize, 37 this result demonstrates that for a small two—cluster network, the SRSA algorithm can closely match the performance of an ideal slot allocation schedule. The performance of SRSA in larger networks is presented in Section 2.4. 2.3.7 Protocol Convergence Convergence and stability of the SRSA protocol can be analyzed using a discrete time Markov chain as shown in Fig. 2.8. A state in this chain represents a combination of slot allocation vectors across all the clusters. And the state identifier k is indicative of the total number of collisions (both CS and HN) in the network as a result of those allocation vectors. The system can be modeled as a memory-less Markov process because the next allocation vector in SRSA depends only on the current collision state, and is not related to previous allocations. In other words, the next system state depends only on the current state. System state transitions are driven by the SRSA flame clock. At any state with non-zero k, SRSA reacts on collision feedback and it forces the allocation vectors to change. As a result, the system state changes. However, once the system enters state-0, SRSA stops changing slot allocations because there are no collisions. This corresponds to a converged stable change. In the analysis in Section 2.3.6, we have observed that if the flame scaling factor is larger than the parameter SF critical, then a collision-flee state (k=0) is guaranteed to exist in the Markov chain in Fig. 2.8. Now, since by the definition of the SRSA protocol syntax, there exists a non-zero transition probability flom state-i to state-j (i = I, 2, ..., N; j = 0, I, N), the system is guaranteed to arrive at the absorbing state 0. Therefore, SRSA will converge as long as the condition SF > SF critical is satisfied. Also, the total 38 time needed for the system to reach from an initial state to state-0 is the protocol convergence time. With higher value of SF, faster convergence is expected. A $®A ______________ Absorbing SIBIGV Fig. 2.8: Model to represent system state with SRSA 2.4 Performance Evaluation SRSA has been simulated using a C-based even-driven sensor network simulator and its performance have been compared with TDMA-over-CDMA (TDCD), TDMA with random allocation (TDRN), and un-slotted l-persistent CSMA. CSMA is chosen to represent the asynchronous MAC protocOls (such as 802.11) for which energy saving is not possible by intermittent interface sleep or idling [16]. In TDRN, independent random slot allocations are done in each cluster, as shown in Fig. 2.4: a. The protocol TDRN can be viewed as SRSA without the iterative allocation reorganization part of it. 2.4.1 Experimental Parameters We simulate a sensor field with 5 overlapping clusters as shown in Fig. 2.1. Clusters have a radius of 50 m, the same as nodes’ transmission range. Each cluster contains 20 arbitrarily placed nodes, each of which can communicate directly with its cluster head. The distance between adjacent cluster-heads is an experimental parameter for studying the effects of cluster overlapping on SRSA performance. For this purpose, we define a parameter, Cluster Overlapping Factor (C 0F) as l — (distance between cluster heads) / (3 "‘ transmission range). With COF = 1, all 5 clusters in Figure 2.1 totally overlap, and 39 with C 0F = 0 (when the adjacent clusters are separated by three times node transmission range), there is no overlapping between any two adjacent clusters. Unless specified otherwise, we have chosen an inter-cluster distance of 70m, for which C 0F = 0.53 and the corresponding sensor field dimension is 240m x 240m. Following the specification in [32], we choose a radio data rate of 25.6 Kbps and fixed packet size of 128 bits (5ms), which is also the TDMA slot duration. Note that this packet duration includes packet data time and carrier-sense duration. TDMA flame duration is chosen as D+N*SF, where D is the number of slots allocated for downlink transmissions, N is the number of nodes per cluster and SF is a chosen flame scaling factor. For all the reported experiments we have chosen1 the value of D as 9. Unless specified otherwise for all experiments we have chosen SF = 2 and N = 20. As for data model, uplink packets are generated flom each sensor node following a Poisson process with varying inter-packet time. We also generate download packets at a constant rate which is enough to saturate the downlink slots. 2.4.2 Simulation Results Throughput: Throughput for all four protocols with Scaling Factor (SF) = 2 and Cluster Overlapping Factor (COF) = 0.53 are shown in Fig.2.8. TDCD (TDMA-over-CDMA), being completely collision flee, can sustain a maximum load of one packet per flame flom each node. With 20 uplink and 1 downlink slots per flame, the flame duration is 105ms (21 x 5ms). So the maximum sustainable load is approximately 9.5 packets/node/sec. Note that TDCD is able to achieve this high ' With K overlapping neighbor clusters, the minimum number of downlink slots for collision flee downlink transmission is 2K + l (K = 4 for the network in Fig. 3.1). 40 throughput at the expense of wide-band operations (CDMA/FDMA) and high cost/complexity of low-cost sensor nodes. H O 1 r ‘ —£—SRSA i 00 1 r Througput (Packets/Node/ Sec) 01234567891011 Traffic Load (Packets/Node/ Sec) Fig. 2.9: Throughput as a function of load Although at steady state, the SRSA protocol does not face any collisions (see Fig. 2.10:a and 2.10:b) its throughput saturates at an earlier load (4.08 packets/node/sec) due to flame scaling. For SRSA too, the maximum sustainable load is one packet per flame flom each node. However, duration of the scaled flame for SRSA is 245ms (9 downlink slots and 40 uplink slots with SF = 2). For TDRN, the flame size is exactly the same as SRSA (i.e. TDRN was run with the same scaling factor of SRSA), but TDRN saturates at a lower load (approximately 3.75 packets/node/sec) because of higher CS and HN collisions as shown in Figs. 2.10:a and 2.10:b. Similarly for CSMA, flequent collisions are responsible for its early throughput saturation at the rate of 3.74 packets/node/sec approximately. Collisions: CS-Collision rate is defined as the average number of transmission 41 attempts aborted due to contested medium, normalized by the successful packet delivery count in a flame. Similarly, HN-Collision rate is computed as the ratio of number of packets dropped due to hidden collisions to the successful packet delivery count. As shown in Fig. 2.10, within the sustainable loading range, both the collisions are almost eliminated for SRSA. Zero HN-Collision rate indicate zero packet drops and zero CS-Collision rate leads to highly reduced (no queuing delay) packet delivery latency. "l 0.8 0.5 o 0.7 o g 0.6 5 0.4 .505 .5 0.3 g 0.4 g L?) 0.3 U? 0.2 8 0.2 g 0.1 0.1 0 0 0 1 2 3 4 0 1 2 3 4 Load (Pkts/Node/S) Load (Pkts/Node/S) a) Impact of load on CS-Collisions b) Impact of load on packet drop Fig. 2.10: CS and HN collisions as a function of load These results indicate the effectiveness of SRSA’s feedback based slot reorganization algorithm in the presence of flame scaling. The effects are particularly visible when contrasted with the TDRN case in which flames are scaled and slots are randomly allocated, but no adaptive reorganization is done for collision control. As expected, for CSMA, both the collisions are higher than both SRSA and TDRN. For TDCD (not shown), there are no collisions within the sustainable loading range of 9.5 42 packets/node/sec. Algorithm Convergence: Convergence characteristics of SRSA on a five-cluster network are plotted in Fig. 2.11. A scaling factor of 2 was used for this experiment. At a load of 1.4 packets per second, with both active and passive CS-Collision detection turned on (see Section 2.3.2), the allocation schedule converges after roughly 400 flames or 98 seconds. For the same set up, turning ofl‘ the passive detection causes the CS-Collision to converge an order of magnitude slower, which is after 5000 flames or 1225 seconds. This clearly indicates the effectiveness of the passive detection component of SRSA. While experimenting with different loading situations, it has been observed that when an optimal allocation exists, SRSA finds it quicker at higher loads. This fact is also evident in Fig. 2.10, in which both CS and EN collision rates have a small non-zero value at lower loads (0.2 through 1 packets/node/sec); but then the collisions are removed completely at higher loads. Better responsiveness of both active and passive CS-Collision detection mechanism at higher loads explains this observation. In the in-band active detection process, a node informs its cluster head about an experienced CS-Collision during the first successful uplink delivery after the collision. At higher loads, the node gets such opportunities sooner compared to at lower loads. As a result, active detection of collisions and their resolution are more effective at higher loads. 43 o 12 5 10 CS-Collisions “‘ 8 ’6 9" 6 E 4 )lN-Collisions o 0 1 i + ‘r 0 50 100 150 200 250 300 350 400 Frame Number a) Collision convergence with both active and passive detection of CS-Collisions 8 , IE 6 CS-Collisions g4 M... I one . 2 go I ll '/ L If % L7 0 1000 211210 301) 4111) 50(1) FraneNurber b) Collision convergence with only active CS—Collision detection Fig. 2.11: Protocol convergence in SRSA Scalabilig: Since the SRSA algorithm uses only local feedback to reallocate slots, it scales well with increasing network size. Fig. 2.12 shows protocol convergence with difl‘erent network sizes. Protocol convergence is depicted in terms of system wide CS-Collisions, when both active and passive collision detections are used with a scaling factor of 3. The first observation is that for 5-cluster case the system converges much faster than that in Fig. 2.11. A higher SF value (3 as opposed to 2 in Fig. 2.11) is responsible for this faster convergence. The second observation is that with larger networlm (9 and 16 clusters) the protocol converges within reasonable time duration of at most 150 SRSA flames. This experiment confirms that although it is little slower, SRSA does converge for larger networks. We have found very similar behavior of HN-Collision patterns with varying network size. 25 N O I 16 clusters u—n M i 5 clusters 9 clusters —- O CS_Collision Per Frame . l 31 61 91 121 Frame Number Fig. 2.12: Protocol convergence for varying network size Coflion Distribtm: The spatial dish-ibution of collision rates is shown in Fig. 2.13. For clarity, collisions for only the center cluster (see Fig. 2.1) are shown. The shown area corresponds to a 100m x 100m square which contains the center circular cluster of a radius of 50m. Each column represents the average packet drop rate (HN-Collision rates) experienced by one of the 20 nodes in that cluster. Observe that for TDRN, the nodes along the periphery, which are located in cluster-overlapped areas, experience more packet drops compared to the nodes near the cluster center. This is because they are more susceptible to HN-Collisions due to their relative location. This trend is preserved even after SRSA’s allocation reorganization algorithm is able to significantly reduce the overall drop rates. This particular experiment was run for a larger cluster overlapping (COF = 0.66), which is why, unlike the last experiments, SRSA was not able to completely remove the HN-Collisions. 45 Spatial Distribution of HN-Collisions in the Center Cluster Packet Drop Rate with TDRN Packet Drop Rate with SRSA Fig. 2.13: Spatial distribution of packet drops I'M: For all evaluated protocols in TDMA family (TDCD, SRSA and TDRD), the MAC latency has two components, namely, the queuing delay2 and a delay imposed by the TDMA flaming itself. The latter is load independent and its average value is half the flame duration. CSMA, being not a flame based protocol, does not have this flame-latency overhead. As a result, as shown in Fig. 2.14, the latency for CSMA is the lowest (1.2 slots) among all four protocols. The latency for CSMA increases marginally with load, but it is not noticeable due to large linear scale in Fig. 2.14. On the other hand, TDCD, with a flame size of 21 slots, has a flame-latency of 10.5 slots. Although there are no collisions for TDCD, with Poisson traffic there is slight queuing which pushes the latency to a range 11.8 slots to 15.4 slots for the shown loading conditions. SRSA has a larger frame size of 49 slots. Therefore its average flame-latency is 24.5 slots. Additional queuing delay increases the latency further for 2 For all our experiments, we have used a fixed buffer size of 500 packets in each sensor node. 46 SRSA. For TDRN, the flame-latency is the same as SRSA because it also has 49 slot flames. However, the queuing delay for TDRN is larger because of more CS-Collisions (see Fig. 2.10) in the absence of allocation reorganization. 250 +SRSA -o—TDRN +CSMA +TDCD mg) 8 Latency (slot ti 3 8 8 A A A A A A A A v v V V V V ' A A A A A A A v V V'v V V V l 2 Traffic Load (Packets/NoddSec) Fig. 2.14: Packet delivery latency as a function of load Enemy Eficz’engg: Energy performance for all four protocols with varying inter-cluster distance is shown in Fig. 2.15. For these experiments we chose a load of 1.38 packets/node/sec. Energy consumption for a protocol is represented by the average number of slots during which a transmitter and a receiver node have to remain active for each successful packet delivery. For the inactive duration, nodes sleep to conserve energy3. Larger node active times indicate higher energy consumption. For all protocols except TDCD, the energy consumption increases with denser cluster placements. CSMA shows the worst energy performance because all nodes have to always remain active in order to receive asynchronous transmissions. On the other hand, for TDMA family of protocols, all nodes except the cluster heads has to be active 3 We assume that a transmitter has to remain active only during 10% of a slot time to detect a CS-Collision. If one is detected, the node can still sleep for the remaining 90% of the slot. 47 only during their respective allocated slots only if there is a packet to transmit. Energy efficiency of TDMA based protocols is evident flom their order of magnitude energy economy (see Fig. 2.15) compared to CSMA. 250 I CSMA 200 W M 150 3 13 8 On a A. 0 .§ [— D .2 6 <2 0 0.2 0.4 0.6 0.8 1 Cluster Overlapping Factor Fig. 2.15: Energy expenditure as node uptime Due to the absence of any collisions, the energy drainage in TDCD is insensitive to cluster overlapping. For SRSA and TDRN, both CS and HN collisions start increasing for larger COFs (see Fig. 2.16). In closely overlapped clusters, in fact there are no optimal allocations by SRSA that can achieve zero CS and HN collisions. In such situations, SRSA attempts to minimize the HN—Collisions at the expense of additional CS-Collisions. Energy wastage due to these collisions (see Fig. 2.10) accounts for the increase in node active time for SRSA. 48 0.7 ‘ . . D ‘5 0.6 ‘ "‘ o 8‘ 52' 0.5 - :6; s: 8 .3 0.4 — 5; =1 c: (3 0.3 ~ .g 8 E . o 0.2 L.) 0.1 — § 0 ‘ 0 :4. ¢ : h a. . r 0 0.2 0.4 0.6 0.8 0 02 0. 0.6 03 Cluster Overlappmg Factor Cluster Overlapping Factor a) Impact of cluster b) Impact of cluster overlapping overlapping on CS-Collisions on packet drop rates Fig. 2.16: Collisions for different cluster overlapping gm of Clu_.ger Overlapping: As evident flom Fig. 2.16, since SRSA suffers flom lower collisions compared to TDRN, the former delivers little bit better energy performance. Overall, as expected, SRSA’s energy performance is significantly better than contention based protocols such as CSMA and is identical to that of TDCD for COF up to 0.4. Beyond that, it increases by approximately 50% when the clusters are almost completely overlapped. Finally, throughput reductions, for CSMA, TDRN and SRSA protocols, due to increased collisions for higher cluster overlapping are shown in Fig. 2.17. Eflecps of Frame Scalmg: As shown in Fig. 2.18, for cluster overlapping factors larger than 0.53, there are no slot allocations that can completely eliminate the HN and CS Collisions for SF = 2. To eliminate collisions in closely overlapped clusters, SRSA is required to use larger flame scaling factors. Fig. 2.18 depicts the rate of both types of 49 collisions with changing overlapping factor for SF = 2, 3, 4 and 5. For example, while the maximum sustainable (zero collisions) overlapping factor is 0.533 with SF = 2, it can be extended to 0.6 with SF = 3, 0.66 with SF = 4, and 0.733 with SF = 5. For higher values of SF, more flee slots are introduced in the TDMA flame. As a result, the SRSA slot reorganization algorithm is given more flexibility to remove collisions by swapping the collision slots with those flee slots. Note that the tradeoffs for higher SF are increased latency and lower effective channel capacity. Conceptually, flame scaling serves the same purpose as that of active-sleep cycles in SMAC [l6] and preamble sampling [32, 33], but in the context of reducing inter-cluster MAC interference. The flee slots in sealed flames extend the effective sleep duration of a node at the expense of lowering allowable traffic loading range. 1.4 T g 1.3 ~~ Q 0 '8 E; 1.2 .- .3. EL 33. 1.1 -- ; g? +SRSA 8 —o—TDRN E 1 fl +CSMA —o—TDCD 0.9 l .1 l t . 0 0.2 0.4 0.6 0.8 1 Cluster Overlapping Factor Fig. 2.17: Throughput with varying cluster overlapping 50 U) O N M r (A '1'] H W O N 1 + Cl) '1'] ‘II VI HN-Collision Rate .o 3 ~— u: 0 0.2 0.4 0.6 0.8 0 0.2 0.4 0.6 0.8 COF COF a) Impact of SF on CS—Collisions b) Impact of SF on packet drop rate Fig. 2.18: Effects of flame scaling on collisions in SRSA Effects of Mobility: The effects of limited node mobility on the SRSA protocol performance are shown in Fig. 2.19. In addition to the regular 100 static nodes, 5 mobile sensor nodes are added in the sensor field in Fig. 2.1. Mobility for those nodes is implemented based on random way-point model, according to which a node randomly chooses a destination and moves towards it with a constant speed (2 m/sec). The degree of mobility is controlled by varying the interval between movements normalized by flame duration. Cluster overlapping factor, traffic load and scaling factors are kept same as those used previously. When a mobile node moves across clusters, the old cluster de-allocates its slot, and the new cluster allocates it a slot flom its pool of flee slots. These new allocations give rise to temporary inter-cluster CS and HN collisions. As evident flom Fig. 2.19, with longer intervals between moves, the allocation reorganization mechanism of SRSA can limit these collisions to a great extent. However, with very short intervals, the mobile nodes move more frequently than the convergence time constant of SRSA. As a result, 51 the collisions are higher. As the interval between moves increases to very large values, it is expected that collisions will converge to zero. In efl‘ect, limited node mobility introduces random perturbation, which increases the convergence time of SRSA. 0-011 + CS-CoIlision ‘ " l _, ,+. . - , 19’5“”??? <_ .. 0.007 - ollision Rate 0.005 i O 0.003 ‘ 0.001 0 200 400 600 800 1000 Interval between moves (Frames) Fig. 2.19: Effects of mobility on Collisions in SRSA However, the absolute collision rates are quite small. With no pause between moves, the drop rate (HN-Collision) is less than 0.5% and the CS-Collision rate is less than 1.2%. These numbers should be compared with the complete static scenario, in which both collisions are nonexistent for SRSA (see Fig. 2.9). Based on the low CS-Collision and drop numbers, we conclude that SRSA scales well for networks with moderate mobility for a small flaction of the sensor nodes. Overall Comparison: From performance summary in Table 2, it is evident that because of its largest energy overhead, CSMA cannot be used for energy-constrained sensor nodes. Among the TDMA based protocols, although TDCD is the best performing but its dependency on CDMA (or FDMA) makes it prohibitive for 52 narrow-band operations with cost-constrained sensor nodes. Between SRSA and TDRN, SRSA is a better choice because of its smaller packet drop rates and lower delivery latency (see Fig. 2.9). In fact, for low traffic loads (which is the case for many monitoring based sensor applications) and moderate cluster overlapping, SRSA can completely eliminate packet dr0ps at the expense of additional latency, which is the only trade-off for this protocol. As a result, SRSA makes an excellent medium access strategy in sensor network applications that can tolerate relatively larger delay but not flequent packet drops. Narrow-band Cost, Through- Packet Latency Energy Operation Complexity put Drop Rate Usage TDCD No @h High VeryLow Low Low SRSA Yes Low Med. Very Low Med. Low TDRN Yes Low Med. Med. Med. Low CSMA Yes Low Med. High Low High ’ Table 2.2: Overall performance summary 2.5 Summary In this work, we have proposed and evaluated a Self-Reorganizing Slot Allocation (SRSA) mechanism for TDMA based Medium Access Control (MAC) in wireless sensor networks. SRSA uses a feedback based local allocation reorganization that can be used for minimizing allocation interference across overlapping sensor clusters in a single channel network. One major advantage of SRSA protocol is that it does not require any global or neighborhood synchronization, making it feasible in very large sensor networks. For moderately dense clusters, SRSA uses a novel flame scaling technique for achieving lower packet drops by reducing the maximum allowable system loading. Comparative evaluation shows that like plain TDMA protocols, SRSA 53 has an order of magnitude lower energy overhead compared to CSMA and other contention based protocols. The trade-off for very low energy dissipation and packet drop rate is increased transmission latency. Because of its independence flom wideband techniques such as CDMA or FDMA, SRSA offers a cost-effective solution for complexity-constrained sensor hardware. To summarize, SRSA makes an ideal medium access strategy in sensor network applications that can tolerate relatively larger delay but not flequent packet drops. Ongoing work on SRSA includes extending the protocol analysis, as outlined in Section 2.3.7, for analytically estimating the convergence speed. Computing the convergence speed according to this model will involve the following. First compute the steady state probability of finding the system in any state-i for all i 76 0. Then evaluate the average number of transitions required flom any state-i to state-0 and then compute the expected value over all for all i 36 0. And we have assumed a reliable TDMA schedule transfer mechanism flom cluster heads to individual sensor nodes. The impacts of errors in such schedule transfer need be investigated. Effects of variable node density across clusters and their effects on variable cluster-specific Scaling Factors (SF) are also being investigated. 54 Chapter 3 Scalable Routing for Very Large Sensor Network This chapter presents an architectural solution to address the problem of scalable routing in very large sensor networks. The control complexities of the existing sensor routing protocols, both node-centric and data-centric, do not scale very well for large networks with potentially hundreds of thousands of embedded sensor devices. We develop a routing solution Off-Network Control Processing (ONCP) that achieves control scalability in large sensor networks by shifting a certain amount of routing functions to an “off-network” server. A tiered and hybrid routing approach, consisting of “coarse grain” server based global routing, and distributed “fine grain” local routing is proposed for achieving scalability by avoiding the network-wide control message dissemination. Section 3.1 describes the related routing protocols for wireless sensor networks. In Section 3.2, we discuss the motivations and problems for scalable routing protocols. We present the network and application model in Section 3.3. In Section 3.4, we discuss the details of the ONCP routing architecture. 3.1 Related Work on Sensor Routing Protocols Compared with wireless ad-hoc networks, routing protocols in sensor networks are particularly challenging due to several unique of its characteristics. (1) It is not possible to build or organize the global addressing for a large number of sensor nodes. Thus classical IP-based protocols cannot be directly applied to sensor networks. (2) Unlike the typical multiple to multiple communication traffic models, sensor networks 55 generally require the data flow flom multiple sources to one (or few) data sink. (3) The generated data has significant redundancy since multiple sensors may generate the similar data within the vicinity of a common phenomenon. These data needs to be processed in-network by the routing protocols in order to improve energy efficiency and bandwidth utilization. (4) The sensor networks comprise a large number of unattended nodes. Depending on specific applications, the network size may be thousands or even up to hundreds of thousands. Scalability in routing protocol is a challenging problem in such large networks. (5) The sensor nodes are tightly constrained in terms of processing and storage capacity, on-board energy supply, wireless communication capacity, etc. All these need to be considered in the protocol design. Many new routing algorithms have been proposed for sensor networks. These mechanisms have considered the characteristics of sensor nodes outlined above, along with the application and architectural requirements. Generally, the routing protocols can be categorized into flat, hierarchical and geographical routing protocols. Most flat protocols are data-centric, in which the routing algorithms rely on the naming of data. The naming mechanisms help to reduce data redundancy. Hierarchical protocols allow only a small set of nodes to participate in routing path discovery and perform data aggregation by organizing nodes in groups and assigning different functionalities. Geographical protocols use the node position information to relay the data rather than flooding routing discovery messages to the whole network. In this section, we will mainly discuss these three classes of routing protocols for wireless sensor networks. 56 3.1.1 Flat Sensor Routing Protocols In Directed Diflhsion [3 8], the sink nodes disseminate network—wide sensing interests for the target sensor nodes. The intermediate nodes propagate interests to establish routing gradients. When the sink nodes receive exploratory data flom target nodes, they eventually reinforce one or multiple gradients along the reverse paths for steady data collection. The energy benefits of Directed Diffusion come flom in-network data processing and empirically selecting desirable paths. However, the protocol suffers flom severe scalability issues caused by the network-wide interest and exploratory data disseminations during the route set up phase. With flequently changing and/or short-lived sensing tasks, the control congestion and energy penalties make this protocol not scale for large sensor networks. Gradient-Based Routing (GBR) [54], Energy Aware Routing [55] and COUGAR [3] are three other flat routing approaches. GBR and Energy Aware Routing employ different optimization techniques flom Directed Diffusion. GBR utilizes the information of hop counts when the interest is disseminated through the network. And Energy Aware Routing maintains and chooses paths by a certain probability to optimize energy usage. COUGAR regards the network as a distributed database system. It uses declarative queries to abstract query processing flom the network layer functions such as selection of relevant sensors as lead nodes to perform aggregation and transmission. All these flat approaches suffer flom the problem of network-wide route searching. In large networks, the overheads of routing discovery will overwhelm their optimization techniques. 57 3.1.2 Hierarchical Sensor Routing Protocols LEACH [5], one of the first hierarchical sensor routing algorithms, proposes to organize nodes into clusters, and only use cluster-heads to communicate directly with base stations. Cluster-heads are periodically rotated in order to balance the energy dissipation. LEACH is completely distributed and its major benefits come flom data fusion at the cluster-heads. The primary shortcoming of LEACH architecture is its assumption that each sensor node is able to make single-hop transmissions to the base stations. For large networks with distant base stations, this will require prohibitively large transmission power which may be impractical for tiny sensor nodes with energy-limited transceivers. PEGASIS [56] is an enhancement over LEACH protocol. In order to extend network lifetime, nodes need to only communicate with their closest neighbors and they take turns in communicating with the base-stations. This reduces the power required to transmit data per round as the power draining is spread uniformly over all nodes. The main drawback of PEGASIS is that it still needs direct nodal communication to base stations. HPAR [57] is a hierarchical power-aware routing protocol. Groups of nodes are organized together as a zone and treated as an entity in routing. Each zone uses max-min algorithm to decide how to route a message hierarchically across neighbor zones such that network life is maximized. Hierarchical protocols help to alleviate the problem of routing flooding messages by only allowing higher layer hierarchy to participate in routing path establishment. Yet they still can not avoid network-wide 58 searching, which can be severely expensive in large networks. And the overheads for establishing network hierarchy also restrict their efficiency. 3.1.3 Geographic Sensor Routing Protocols The routing protocol GEAR [41] partially addresses the interest flooding problem of Directed Diflusion by geographically routing the interests towards the source nodes specified by the data attributes in interest packets. The key routing idea for a node is to always forward interest packets to its neighbors closer to the physical location of the source nodes. When this basic forwarding fails due to the fact that the forwarding node itself is the geographically closest but not within the radio range of the source region, in GEAR the interest packets are forwarded using cost based detour mechanisms. Other geographical routing protocols such as GPSR [42] and GOAFR [43] follow similar routing philosophies, but differ in the detour search process. While these geographical routing protocols provide better control message complexity than Directed Diffusion, the assumption of complete node localization makes them less feasible for localization-denied sensor fields. Also, in the event of flequent detour searches, the additional control overhead lessens the relative advantages of these protocols over Directed Diffusion. Frequent detour searches may be caused by irregular terrains and other security based partitioning. Another family of energy aware routing protocols [49, 50, 51] is proposed to compute the optimal routing schemes that maximize the life time of networks. These are formulated as a class of optimization problem and heuristics are applied to select energy-efficient routes in a distributed manner. The primary limitation of these 59 approaches is that they require prior knowledge of the source and destination sensor pairs and the exact traffrc pattern. 3.2 Routing Scalability: Motivation and Problem Description Wireless sensor networking is an emerging enabling technology for a wide range of embedded applications such as large scale environmental monitoring, intrusion detection, and collaborative disaster management. The sensor population for such applications can often be very large due to the expanse of the targeted sensor fields and the limited sensing and radio range of the sensor nodes. For example, the networks used for environmental monitoring are likely to be deployed over tens of square miles of target area comprising of potentially hundreds of thousands sensor nodes. We focus on a major design constraint for very large sensor networks: routing scalability. To establish routes, protocols generally incur network wide control message flooding. For traditional node-centric routing, routing messages are generated primarily for topology discovery in link-state protocols such as OLSR [39], and for route discovery in distance-vector protocols such as AODV [40]. In data-centric protocols such as Directed Diffusion [3 8], the control message complexity manifests in the form of network wide interest and exploratory data propagation during the routing discovery process. For flequent and short-lived sensing tasks, the control overhead can be overwhelming in very large sensor networks (VLSN). High control overhead can have a series of adverse effects including high packet drop rate, increased latency, and reduced effective network bandwidth. A more detrimental effect is the increased energy overhead on the resource constrained sensor nodes. 60 To address this problem, we prOpose an architectural solution by developing an Off-Network Control Processing (ONCP) flamework for Very Large Sensor Networks. The key idea is to shift a significant amount of routing functionality “off-network” to an energy-unlimited server. This server performs “coarse grain” global route computation, while the “fine grain” local routes are discovered solely by local sensor nodes. This partially server-based approach allows eliminating the control messages that otherwise are needed for distributed global route establishment. The reduction of routing control messages by the ONCP server ensures routing scalability in large networks. 3.3 Network and Application Model As depicted in Fig. 3.1, our network model consists of multiple base stations which act as data sinks. The base stations are connected to an Off-Network Control Processing (ONCP) server through a low-latency wired network, which is used for network status update to the ONCP server, and for global route download flom the ONCP to the sensor network. As for the application model, each user is connected to the ONCP server through a client that generates asynchronous sensing task requests with specific geographical or other attributes. It is envisioned that a user client will include a Graphical User Interface (GUI) that renders a terrain map of the sensor field with environmental features such as vegetation, mountains, water bodies, and other similar landmarks. A user will first identify a target sensing area through the client GUI and then generate sensing task requests to the ONCP server. For example, a user may select a certain part of a 61 vegetation area as the target sensing zone, and then deliver a sensing task request to measure the soil pH level at a regular interval of six hours and for duration of thirty days. Based on their target sensing areas, the task requests are routed to the appropriate sensor nodes. Note that a user does not address or specify the sensor nodes and therefore, no individual sensor level information is made available to the users. User-1 User-2 User-n ONCP . ‘ I .-.r ‘."'j.-‘.z.‘.'.‘..~:o.-. . :1 9.1:”. p.13; 9:557.“ Ar ' {‘23}, H-f‘f-K if, 74:2:fls1..»,_‘,.‘ Lu 59?” Off Work cartoon “ pm “ ” ~' 9.9 ' Data/Events Commands Base Sensors Field fl : 35’3 Station-1 5k Fig. 3.1: Network and application model 3.4 Off-Network Control Processing (ONCP) Routing Architecture ONCP is a hybrid architecture in which the routing decisions are made at two levels. Contrary to the traditional hierarchical sensor routing protocols [30], the global path selection in ONCP is performed solely in a centralized server, which does not require network-wide route search. And the local routing messages are constrained only locally. For large networks, this hybrid route discovery mechanism achieves excellent control plane scalability compared to the flat and hierarchical protocols described in Section 3.1. 62 As shown in Fig. 3.2, the global routes are computed by the ONCP server at the level of pre-configured regions. When a user client generates a sensing task to certain sensors in a target source region, the ONCP server computes an energy—optimal global route flom the target source region to an appropriate base station. Such a coarse-grain global route is represented by specifying a sequence of regions fl'om the target source region to the chosen base station. To realize this region level route, each region on the route maintains a next-region routing entry for each individual sensing task. At the lower tier, ONCP implements a fine-grain reactive local routing protocol that provides physical routing paths at node level across adjacent regions. The local routes are established in a completely local and distributed manner. A clustering abstraction has been also used for achieving scalability for the local routing. 1 Region 1A A Global Routes ‘ I ' ....... . ......... é)°/‘\ A; A A kc" ‘ A ;/:'-\"J’:\/€‘ - / A - 7 féR‘egions: Al\‘ ‘ gRegionAB V -A Region 4 Regiox 7 Region B/r—A