diff options
Diffstat (limited to 'doc/rfc/rfc5867.txt')
-rw-r--r-- | doc/rfc/rfc5867.txt | 1459 |
1 files changed, 1459 insertions, 0 deletions
diff --git a/doc/rfc/rfc5867.txt b/doc/rfc/rfc5867.txt new file mode 100644 index 0000000..f2446fd --- /dev/null +++ b/doc/rfc/rfc5867.txt @@ -0,0 +1,1459 @@ + + + + + + +Internet Engineering Task Force (IETF) J. Martocci, Ed. +Request for Comments: 5867 Johnson Controls Inc. +Category: Informational P. De Mil +ISSN: 2070-1721 Ghent University - IBCN + N. Riou + Schneider Electric + W. Vermeylen + Arts Centre Vooruit + June 2010 + + + Building Automation Routing Requirements + in Low-Power and Lossy Networks + +Abstract + + The Routing Over Low-Power and Lossy (ROLL) networks Working Group + has been chartered to work on routing solutions for Low-Power and + Lossy Networks (LLNs) in various markets: industrial, commercial + (building), home, and urban networks. Pursuant to this effort, this + document defines the IPv6 routing requirements for building + automation. + +Status of This Memo + + This document is not an Internet Standards Track specification; it is + published for informational purposes. + + This document is a product of the Internet Engineering Task Force + (IETF). It represents the consensus of the IETF community. It has + received public review and has been approved for publication by the + Internet Engineering Steering Group (IESG). Not all documents + approved by the IESG are a candidate for any level of Internet + Standard; see Section 2 of RFC 5741. + + Information about the current status of this document, any errata, + and how to provide feedback on it may be obtained at + http://www.rfc-editor.org/info/rfc5867. + + + + + + + + + + + + + +Martocci, et al. Informational [Page 1] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +Copyright Notice + + Copyright (c) 2010 IETF Trust and the persons identified as the + document authors. All rights reserved. + + This document is subject to BCP 78 and the IETF Trust's Legal + Provisions Relating to IETF Documents + (http://trustee.ietf.org/license-info) in effect on the date of + publication of this document. Please review these documents + carefully, as they describe your rights and restrictions with respect + to this document. Code Components extracted from this document must + include Simplified BSD License text as described in Section 4.e of + the Trust Legal Provisions and are provided without warranty as + described in the Simplified BSD License. + + This document may contain material from IETF Documents or IETF + Contributions published or made publicly available before November + 10, 2008. The person(s) controlling the copyright in some of this + material may not have granted the IETF Trust the right to allow + modifications of such material outside the IETF Standards Process. + Without obtaining an adequate license from the person(s) controlling + the copyright in such materials, this document may not be modified + outside the IETF Standards Process, and derivative works of it may + not be created outside the IETF Standards Process, except to format + it for publication as an RFC or to translate it into languages other + than English. + + + + + + + + + + + + + + + + + + + + + + + + + +Martocci, et al. Informational [Page 2] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +Table of Contents + + 1. Introduction ....................................................4 + 2. Terminology .....................................................6 + 2.1. Requirements Language ......................................6 + 3. Overview of Building Automation Networks ........................6 + 3.1. Introduction ...............................................6 + 3.2. Building Systems Equipment .................................7 + 3.2.1. Sensors/Actuators ...................................7 + 3.2.2. Area Controllers ....................................7 + 3.2.3. Zone Controllers ....................................8 + 3.3. Equipment Installation Methods .............................8 + 3.4. Device Density .............................................9 + 3.4.1. HVAC Device Density .................................9 + 3.4.2. Fire Device Density .................................9 + 3.4.3. Lighting Device Density ............................10 + 3.4.4. Physical Security Device Density ...................10 + 4. Traffic Pattern ................................................10 + 5. Building Automation Routing Requirements .......................12 + 5.1. Device and Network Commissioning ..........................12 + 5.1.1. Zero-Configuration Installation ....................12 + 5.1.2. Local Testing ......................................12 + 5.1.3. Device Replacement .................................13 + 5.2. Scalability ...............................................13 + 5.2.1. Network Domain .....................................13 + 5.2.2. Peer-to-Peer Communication .........................13 + 5.3. Mobility ..................................................13 + 5.3.1. Mobile Device Requirements .........................14 + 5.4. Resource Constrained Devices ..............................15 + 5.4.1. Limited Memory Footprint on Host Devices ...........15 + 5.4.2. Limited Processing Power for Routers ...............15 + 5.4.3. Sleeping Devices ...................................15 + 5.5. Addressing ................................................16 + 5.6. Manageability .............................................16 + 5.6.1. Diagnostics ........................................17 + 5.6.2. Route Tracking .....................................17 + 5.7. Route Selection ...........................................17 + 5.7.1. Route Cost .........................................17 + 5.7.2. Route Adaptation ...................................18 + 5.7.3. Route Redundancy ...................................18 + 5.7.4. Route Discovery Time ...............................18 + 5.7.5. Route Preference ...................................18 + 5.7.6. Real-Time Performance Measures .....................18 + 5.7.7. Prioritized Routing ................................18 + + + + + + + +Martocci, et al. Informational [Page 3] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + 5.8. Security Requirements .....................................19 + 5.8.1. Building Security Use Case .........................19 + 5.8.2. Authentication .....................................20 + 5.8.3. Encryption .........................................20 + 5.8.4. Disparate Security Policies ........................21 + 5.8.5. Routing Security Policies to Sleeping Devices ......21 + 6. Security Considerations ........................................21 + 7. Acknowledgments ................................................22 + 8. References .....................................................22 + 8.1. Normative References ......................................22 + 8.2. Informative References ....................................22 + Appendix A. Additional Building Requirements ......................23 + A.1. Additional Commercial Product Requirements ................23 + A.1.1. Wired and Wireless Implementations .................23 + A.1.2. World-Wide Applicability ...........................23 + A.2. Additional Installation and Commissioning Requirements ....23 + A.2.1. Unavailability of an IP Network ....................23 + A.3. Additional Network Requirements ...........................23 + A.3.1. TCP/UDP ............................................23 + A.3.2. Interference Mitigation ............................23 + A.3.3. Packet Reliability .................................24 + A.3.4. Merging Commissioned Islands .......................24 + A.3.5. Adjustable Routing Table Sizes .....................24 + A.3.6. Automatic Gain Control .............................24 + A.3.7. Device and Network Integrity .......................24 + A.4. Additional Performance Requirements .......................24 + A.4.1. Data Rate Performance ..............................24 + A.4.2. Firmware Upgrades ..................................25 + A.4.3. Route Persistence ..................................25 + +1. Introduction + + The Routing Over Low-Power and Lossy (ROLL) networks Working Group + has been chartered to work on routing solutions for Low-Power and + Lossy Networks (LLNs) in various markets: industrial, commercial + (building), home, and urban networks. Pursuant to this effort, this + document defines the IPv6 routing requirements for building + automation. + + Commercial buildings have been fitted with pneumatic, and + subsequently electronic, communication routes connecting sensors to + their controllers for over one hundred years. Recent economic and + technical advances in wireless communication allow facilities to + increasingly utilize a wireless solution in lieu of a wired solution, + thereby reducing installation costs while maintaining highly reliant + communication. + + + + + +Martocci, et al. Informational [Page 4] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + The cost benefits and ease of installation of wireless sensors allow + customers to further instrument their facilities with additional + sensors, providing tighter control while yielding increased energy + savings. + + Wireless solutions will be adapted from their existing wired + counterparts in many of the building applications including, but not + limited to, heating, ventilation, and air conditioning (HVAC); + lighting; physical security; fire; and elevator/lift systems. These + devices will be developed to reduce installation costs while + increasing installation and retrofit flexibility, as well as + increasing the sensing fidelity to improve efficiency and building + service quality. + + Sensing devices may be battery-less, battery-powered, or mains- + powered. Actuators and area controllers will be mains-powered. Due + to building code and/or device density (e.g., equipment room), it is + envisioned that a mix of wired and wireless sensors and actuators + will be deployed within a building. + + Building management systems (BMSs) are deployed in a large set of + vertical markets including universities, hospitals, government + facilities, kindergarten through high school (K-12), pharmaceutical + manufacturing facilities, and single-tenant or multi-tenant office + buildings. These buildings range in size from 100K-sq.-ft. + structures (5-story office buildings), to 1M-sq.-ft. skyscrapers + (100-story skyscrapers), to complex government facilities such as the + Pentagon. The described topology is meant to be the model to be used + in all of these types of environments but clearly must be tailored to + the building class, building tenant, and vertical market being + served. + + Section 3 describes the necessary background to understand the + context of building automation including the sensor, actuator, area + controller, and zone controller layers of the topology; typical + device density; and installation practices. + + Section 4 defines the traffic flow of the aforementioned sensors, + actuators, and controllers in commercial buildings. + + Section 5 defines the full set of IPv6 routing requirements for + commercial buildings. + + + + + + + + + +Martocci, et al. Informational [Page 5] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + Appendix A documents important commercial building requirements that + are out of scope for routing yet will be essential to the final + acceptance of the protocols used within the building. + + Section 3 and Appendix A are mainly included for educational + purposes. + + The expressed aim of this document is to provide the set of IPv6 + routing requirements for LLNs in buildings, as described in + Section 5. + +2. Terminology + + For a description of the terminology used in this specification, + please see [ROLL-TERM]. + +2.1. Requirements Language + + The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", + "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this + document are to be interpreted as described in [RFC2119]. + +3. Overview of Building Automation Networks + +3.1. Introduction + + To understand the network systems requirements of a building + management system in a commercial building, this document uses a + framework to describe the basic functions and composition of the + system. A BMS is a hierarchical system of sensors, actuators, + controllers, and user interface devices that interoperate to provide + a safe and comfortable environment while constraining energy costs. + + A BMS is divided functionally across different but interrelated + building subsystems such as heating, ventilation, and air + conditioning (HVAC); fire; security; lighting; shutters; and + elevator/lift control systems, as denoted in Figure 1. + + Much of the makeup of a BMS is optional and installed at the behest + of the customer. Sensors and actuators have no standalone + functionality. All other devices support partial or complete + standalone functionality. These devices can optionally be tethered + to form a more cohesive system. The customer requirements dictate + the level of integration within the facility. This architecture + provides excellent fault tolerance since each node is designed to + operate in an independent mode if the higher layers are unavailable. + + + + + +Martocci, et al. Informational [Page 6] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + +------+ +-----+ +------+ +------+ +------+ +------+ + + Bldg App'ns | | | | | | | | | | | | + + | | | | | | | | | | | | + + Building Cntl | | | | | S | | L | | S | | E | + + | | | | | E | | I | | H | | L | + + Area Control | H | | F | | C | | G | | U | | E | + + | V | | I | | U | | H | | T | | V | + + Zone Control | A | | R | | R | | T | | T | | A | + + | C | | E | | I | | I | | E | | T | + + Actuators | | | | | T | | N | | R | | O | + + | | | | | Y | | G | | S | | R | + + Sensors | | | | | | | | | | | | + + +------+ +-----+ +------+ +------+ +------+ +------+ + + Figure 1: Building Systems and Devices + +3.2. Building Systems Equipment + +3.2.1. Sensors/Actuators + + As Figure 1 indicates, a BMS may be composed of many functional + stacks or silos that are interoperably woven together via building + applications. Each silo has an array of sensors that monitor the + environment and actuators that modify the environment, as determined + by the upper layers of the BMS topology. The sensors typically are + at the edge of the network structure, providing environmental data + for the system. The actuators are the sensors' counterparts, + modifying the characteristics of the system, based on the sensor data + and the applications deployed. + +3.2.2. Area Controllers + + An area describes a small physical locale within a building, + typically a room. HVAC (temperature and humidity) and lighting (room + lighting, shades, solar loads) vendors oftentimes deploy area + controllers. Area controllers are fed by sensor inputs that monitor + + + +Martocci, et al. Informational [Page 7] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + the environmental conditions within the room. Common sensors found + in many rooms that feed the area controllers include temperature, + occupancy, lighting load, solar load, and relative humidity. Sensors + found in specialized rooms (such as chemistry labs) might include air + flow, pressure, and CO2 and CO particle sensors. Room actuation + includes temperature setpoint, lights, and blinds/curtains. + +3.2.3. Zone Controllers + + Zone controllers support a similar set of characteristics to area + controllers, albeit for an extended space. A zone is normally a + logical grouping or functional division of a commercial building. A + zone may also coincidentally map to a physical locale such as a + floor. + + Zone controllers may have direct sensor inputs (smoke detectors for + fire), controller inputs (room controllers for air handlers in HVAC), + or both (door controllers and tamper sensors for security). Like + area/room controllers, zone controllers are standalone devices that + operate independently or may be attached to the larger network for + more synergistic control. + +3.3. Equipment Installation Methods + + A BMS is installed very differently from most other IT networks. IT + networks are typically installed as an overlay onto the existing + environment and are installed from the inside out. That is, the + network wiring infrastructure is installed; the switches, routers, + and servers are connected and made operational; and finally, the + endpoints (e.g., PCs, VoIP phones) are added. + + BMSs, on the other hand, are installed from the outside in. That is, + the endpoints (thermostats, lights, smoke detectors) are installed in + the spaces first; local control is established in each room and + tested for proper operation. The individual rooms are later lashed + together into a subsystem (e.g., lighting). The individual + subsystems (e.g., lighting, HVAC) then coalesce. Later, the entire + system may be merged onto the enterprise network. + + The rationale for this is partly due to the different construction + trades having access to a building under construction at different + times. The sheer size of a building often dictates that even a + single trade may have multiple independent teams working + simultaneously. Furthermore, the HVAC, lighting, and fire systems + must be fully operational before the building can obtain its + occupancy permit. Hence, the BMS must be in place and configured + well before any of the IT servers (DHCP; Authentication, + Authorization, and Accounting (AAA); DNS; etc.) are operational. + + + +Martocci, et al. Informational [Page 8] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + This implies that the BMS cannot rely on the availability of the IT + network infrastructure or application servers. Rather, the BMS + installation should be planned to dovetail into the IT system once + the IT system is available for easy migration onto the IT network. + Front-end planning of available switch ports, cable runs, access + point (AP) placement, firewalls, and security policies will + facilitate this adoption. + +3.4. Device Density + + Device density differs, depending on the application and as dictated + by the local building code requirements. The following subsections + detail typical installation densities for different applications. + +3.4.1. HVAC Device Density + + HVAC room applications typically have sensors/actuators and + controllers spaced about 50 ft. apart. In most cases, there is a 3:1 + ratio of sensors/actuators to controllers. That is, for each room + there is an installed temperature sensor, flow sensor, and damper + actuator for the associated room controller. + + HVAC equipment room applications are quite different. An air handler + system may have a single controller with up to 25 sensors and + actuators within 50 ft. of the air handler. A chiller or boiler is + also controlled with a single equipment controller instrumented with + 25 sensors and actuators. Each of these devices would be + individually addressed since the devices are mandated or optional as + defined by the specified HVAC application. Air handlers typically + serve one or two floors of the building. Chillers and boilers may be + installed per floor, but many times they service a wing, building, or + the entire complex via a central plant. + + These numbers are typical. In special cases, such as clean rooms, + operating rooms, pharmaceutical facilities, and labs, the ratio of + sensors to controllers can increase by a factor of three. Tenant + installations such as malls would opt for packaged units where much + of the sensing and actuation is integrated into the unit; here, a + single device address would serve the entire unit. + +3.4.2. Fire Device Density + + Fire systems are much more uniformly installed, with smoke detectors + installed about every 50 ft. This is dictated by local building + codes. Fire pull boxes are installed uniformly about every 150 ft. + A fire controller will service a floor or wing. The fireman's fire + panel will service the entire building and typically is installed in + the atrium. + + + +Martocci, et al. Informational [Page 9] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +3.4.3. Lighting Device Density + + Lighting is also very uniformly installed, with ballasts installed + approximately every 10 ft. A lighting panel typically serves 48 to + 64 zones. Wired systems tether many lights together into a single + zone. Wireless systems configure each fixture independently to + increase flexibility and reduce installation costs. + +3.4.4. Physical Security Device Density + + Security systems are non-uniformly oriented, with heavy density near + doors and windows and lighter density in the building's interior + space. + + The recent influx of interior and perimeter camera systems is + increasing the security footprint. These cameras are atypical + endpoints requiring up to 1 megabit/second (Mbit/s) data rates per + camera, as contrasted by the few kbit/s needed by most other BMS + sensing equipment. Previously, camera systems had been deployed on + proprietary wired high-speed networks. More recent implementations + utilize wired or wireless IP cameras integrated into the enterprise + LAN. + +4. Traffic Pattern + + The independent nature of the automation subsystems within a building + can significantly affect network traffic patterns. Much of the real- + time sensor environmental data and actuator control stays within the + local LLN environment, while alarms and other event data will + percolate to higher layers. + + Each sensor in the LLN unicasts point to point (P2P) about 200 bytes + of sensor data to its associated controller each minute and expects + an application acknowledgment unicast returned from the destination. + Each controller unicasts messages at a nominal rate of 6 kbit/minute + to peer or supervisory controllers. Thirty percent of each node's + packets are destined for other nodes within the LLN. Seventy percent + of each node's packets are destined for an aggregation device + (multipoint to point (MP2P)) and routed off the LLN. These messages + also require a unicast acknowledgment from the destination. The + above values assume direct node-to-node communication; meshing and + error retransmissions are not considered. + + Multicasts (point to multipoint (P2MP)) to all nodes in the LLN occur + for node and object discovery when the network is first commissioned. + This data is typically a one-time bind that is henceforth persisted. + Lighting systems will also readily use multicasting during normal + operations to turn banks of lights "on" and "off" simultaneously. + + + +Martocci, et al. Informational [Page 10] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + BMSs may be either polled or event-based. Polled data systems will + generate a uniform and constant packet load on the network. Polled + architectures, however, have proven not to be scalable. Today, most + vendors have developed event-based systems that pass data on event. + These systems are highly scalable and generate low data on the + network at quiescence. Unfortunately, the systems will generate a + heavy load on startup since all initial sensor data must migrate to + the controller level. They also will generate a temporary but heavy + load during firmware upgrades. This latter load can normally be + mitigated by performing these downloads during off-peak hours. + + Devices will also need to reference peers periodically for sensor + data or to coordinate operation across systems. Normally, though, + data will migrate from the sensor level upwards through the local and + area levels, and then to the supervisory level. Traffic bottlenecks + will typically form at the funnel point from the area controllers to + the supervisory controllers. + + Initial system startup after a controlled outage or unexpected power + failure puts tremendous stress on the network and on the routing + algorithms. A BMS is comprised of a myriad of control algorithms at + the room, area, zone, and enterprise layers. When these control + algorithms are at quiescence, the real-time data rate is small, and + the network will not saturate. An overall network traffic load of 6 + kbit/s is typical at quiescence. However, upon any power loss, the + control loops and real-time data quickly atrophy. A short power + disruption of only 10 minutes may have a long-term deleterious impact + on the building control systems, taking many hours to regain proper + control. Control applications that cannot handle this level of + disruption (e.g., hospital operating rooms) must be fitted with a + secondary power source. + + Power disruptions are unexpected and in most cases will immediately + impact lines-powered devices. Power disruptions, however, are + transparent to battery-powered devices. These devices will continue + to attempt to access the LLN during the outage. Battery-powered + devices designed to buffer data that has not been delivered will + further stress network operations when power returns. + + Upon restart, lines-powered devices will naturally dither due to + primary equipment delays or variance in the device self-tests. + However, most lines-powered devices will be ready to access the LLN + network within 10 seconds of power-up. Empirical testing indicates + that routes acquired during startup will tend to be very oblique + since the available neighbor lists are incomplete. This demands an + adaptive routing protocol to allow for route optimization as the + network stabilizes. + + + + +Martocci, et al. Informational [Page 11] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +5. Building Automation Routing Requirements + + Following are the building automation routing requirements for + networks used to integrate building sensor, actuator, and control + products. These requirements are written not presuming any + preordained network topology, physical media (wired), or radio + technology (wireless). + +5.1. Device and Network Commissioning + + Building control systems typically are installed and tested by + electricians having little computer knowledge and no network + communication knowledge whatsoever. These systems are often + installed during the building construction phase, before the drywall + and ceilings are in place. For new construction projects, the + building enterprise IP network is not in place during installation of + the building control system. For retrofit applications, the + installer will still operate independently from the IP network so as + not to affect network operations during the installation phase. + + In traditional wired systems, correct operation of a light + switch/ballast pair was as simple as flipping on the light switch. + In wireless applications, the tradesperson has to assure the same + operation, yet be sure the operation of the light switch is + associated with the proper ballast. + + System-level commissioning will later be deployed using a more + computer savvy person with access to a commissioning device (e.g., a + laptop computer). The completely installed and commissioned + enterprise IP network may or may not be in place at this time. + Following are the installation routing requirements. + +5.1.1. Zero-Configuration Installation + + It MUST be possible to fully commission network devices without + requiring any additional commissioning device (e.g., a laptop). From + the ROLL perspective, "zero configuration" means that a node can + obtain an address and join the network on its own, without human + intervention. + +5.1.2. Local Testing + + During installation, the room sensors, actuators, and controllers + SHOULD be able to route packets amongst themselves and to any other + device within the LLN, without requiring any additional routing + infrastructure or routing configuration. + + + + + +Martocci, et al. Informational [Page 12] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +5.1.3. Device Replacement + + To eliminate the need to reconfigure the application upon replacing a + failed device in the LLN, the replaced device must be able to + advertise the old IP address of the failed device in addition to its + new IP address. The routing protocols MUST support hosts and routers + that advertise multiple IPv6 addresses. + +5.2. Scalability + + Building control systems are designed for facilities from 50,000 sq. + ft. to 1M+ sq. ft. The networks that support these systems must + cost-effectively scale accordingly. In larger facilities, + installation may occur simultaneously on various wings or floors, yet + the end system must seamlessly merge. Following are the scalability + requirements. + +5.2.1. Network Domain + + The routing protocol MUST be able to support networks with at least + 2,000 nodes, where 1,000 nodes would act as routers and the other + 1,000 nodes would be hosts. Subnetworks (e.g., rooms, primary + equipment) within the network must support up to 255 sensors and/or + actuators. + +5.2.2. Peer-to-Peer Communication + + The data domain for commercial BMSs may sprawl across a vast portion + of the physical domain. For example, a chiller may reside in the + facility's basement due to its size, yet the associated cooling + towers will reside on the roof. The cold-water supply and return + pipes snake through all of the intervening floors. The feedback + control loops for these systems require data from across the + facility. + + A network device MUST be able to communicate in an end-to-end manner + with any other device on the network. Thus, the routing protocol + MUST provide routes between arbitrary hosts within the appropriate + administrative domain. + +5.3. Mobility + + Most devices are affixed to walls or installed on ceilings within + buildings. Hence, the mobility requirements for commercial buildings + are few. However, in wireless environments, location tracking of + occupants and assets is gaining favor. Asset-tracking applications, + such as tracking capital equipment (e.g., wheelchairs) in medical + + + + +Martocci, et al. Informational [Page 13] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + facilities, require monitoring movement with granularity of a minute; + however, tracking babies in a pediatric ward would require latencies + less than a few seconds. + + The following subsections document the mobility requirements in the + routing layer for mobile devices. Note, however, that mobility can + be implemented at various layers of the system, and the specific + requirements depend on the chosen layer. For instance, some devices + may not depend on a static IP address and are capable of re- + establishing application-level communications when given a new IP + address. Alternatively, mobile IP may be used, or the set of routers + in a building may give an impression of a building-wide network and + allow devices to retain their addresses regardless of where they are, + handling routing between the devices in the background. + +5.3.1. Mobile Device Requirements + + To minimize network dynamics, mobile devices while in motion should + not be allowed to act as forwarding devices (routers) for other + devices in the LLN. Network configuration should allow devices to be + configured as routers or hosts. + +5.3.1.1. Device Mobility within the LLN + + An LLN typically spans a single floor in a commercial building. + Mobile devices may move within this LLN. For example, a wheelchair + may be moved from one room on the floor to another room on the same + floor. + + A mobile LLN device that moves within the confines of the same LLN + SHOULD re-establish end-to-end communication with a fixed device also + in the LLN within 5 seconds after it ceases movement. The LLN + network convergence time should be less than 10 seconds once the + mobile device stops moving. + +5.3.1.2. Device Mobility across LLNs + + A mobile device may move across LLNs, such as a wheelchair being + moved to a different floor. + + A mobile device that moves outside of its original LLN SHOULD re- + establish end-to-end communication with a fixed device also in the + new LLN within 10 seconds after the mobile device ceases movement. + The network convergence time should be less than 20 seconds once the + mobile device stops moving. + + + + + + +Martocci, et al. Informational [Page 14] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +5.4. Resource Constrained Devices + + Sensing and actuator device processing power and memory may be 4 + orders of magnitude less (i.e., 10,000x) than many more traditional + client devices on an IP network. The routing mechanisms must + therefore be tailored to fit these resource constrained devices. + +5.4.1. Limited Memory Footprint on Host Devices + + The software size requirement for non-routing devices (e.g., sleeping + sensors and actuators) SHOULD be implementable in 8-bit devices with + no more than 128 KB of memory. + +5.4.2. Limited Processing Power for Routers + + The software size requirements for routing devices (e.g., room + controllers) SHOULD be implementable in 8-bit devices with no more + than 256 KB of flash memory. + +5.4.3. Sleeping Devices + + Sensing devices will, in some cases, utilize battery power or energy + harvesting techniques for power and will operate mostly in a sleep + mode to maintain power consumption within a modest budget. The + routing protocol MUST take into account device characteristics such + as power budget. + + Typically, sensor battery life (2,000 mAh) needs to extend for at + least 5 years when the device is transmitting its data (200 octets) + once per minute over a low-power transceiver (25 mA) and expecting an + application acknowledgment. In this case, the transmitting device + must leave its receiver in a high-powered state, awaiting the return + of the application ACK. To minimize this latency, a highly efficient + routing protocol that minimizes hops, and hence end-to-end + communication, is required. The routing protocol MUST take into + account node properties, such as "low-powered node", that produce + efficient low-latency routes that minimize radio "on" time for these + devices. + + Sleeping devices MUST be able to receive inbound data. Messages sent + to battery-powered nodes MUST be buffered by the last-hop router for + a period of at least 20 seconds when the destination node is + currently in its sleep cycle. + + + + + + + + +Martocci, et al. Informational [Page 15] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +5.5. Addressing + + Building management systems require different communication schemes + to solicit or post network information. Multicasts or anycasts need + to be used to decipher unresolved references within a device when the + device first joins the network. + + As with any network communication, multicasting should be minimized. + This is especially a problem for small embedded devices with limited + network bandwidth. Multicasts are typically used for network joins + and application binding in embedded systems. Routing MUST support + anycast, unicast, and multicast. + +5.6. Manageability + + As previously noted in Section 3.3, installation of LLN devices + within a BMS follows an "outside-in" work flow. Edge devices are + installed first and tested for communication and application + integrity. These devices are then aggregated into islands, then + LLNs, and later affixed onto the enterprise network. + + The need for diagnostics most often occurs during the installation + and commissioning phase, although at times diagnostic information may + be requested during normal operation. Battery-powered wireless + devices typically will have a self-diagnostic mode that can be + initiated via a button press on the device. The device will display + its link status and/or end-to-end connectivity when the button is + pressed. Lines-powered devices will continuously display + communication status via a bank of LEDs, possibly denoting signal + strength and end-to-end application connectivity. + + The local diagnostics noted above oftentimes are suitable for + defining room-level networks. However, as these devices aggregate, + system-level diagnostics may need to be executed to ameliorate route + vacillation, excessive hops, communication retries, and/or network + bottlenecks. + + In operational networks, due to the mission-critical nature of the + application, the LLN devices will be temporally monitored by the + higher layers to assure that communication integrity is maintained. + Failure to maintain this communication will result in an alarm being + forwarded to the enterprise network from the monitoring node for + analysis and remediation. + + + + + + + + +Martocci, et al. Informational [Page 16] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + In addition to the initial installation and commissioning of the + system, it is equally important for the ongoing maintenance of the + system to be simple and inexpensive. This implies a straightforward + device swap when a failed device is replaced, as noted in Section + 5.1.3. + +5.6.1. Diagnostics + + To improve diagnostics, the routing protocol SHOULD be able to be + placed in and out of "verbose" mode. Verbose mode is a temporary + debugging mode that provides additional communication information + including, at least, the total number of routed packets sent and + received, the number of routing failures (no route available), + neighbor table members, and routing table entries. The data provided + in verbose mode should be sufficient that a network connection graph + could be constructed and maintained by the monitoring node. + + Diagnostic data should be kept by the routers continuously and be + available for solicitation at any time by any other node on the + internetwork. Verbose mode will be activated/deactivated via + unicast, multicast, or other means. Devices having available + resources may elect to support verbose mode continuously. + +5.6.2. Route Tracking + + Route diagnostics SHOULD be supported, providing information such as + route quality, number of hops, and available alternate active routes + with associated costs. Route quality is the relative measure of + "goodness" of the selected source to destination route as compared to + alternate routes. This composite value may be measured as a function + of hop count, signal strength, available power, existing active + routes, or any other criteria deemed by ROLL as the route cost + differentiator. + +5.7. Route Selection + + Route selection determines reliability and quality of the + communication among the devices by optimizing routes over time and + resolving any nuances developed at system startup when nodes are + asynchronously adding themselves to the network. + +5.7.1. Route Cost + + The routing protocol MUST support a metric of route quality and + optimize selection according to such metrics within constraints + established for links along the routes. These metrics SHOULD reflect + metrics such as signal strength, available bandwidth, hop count, + energy availability, and communication error rates. + + + +Martocci, et al. Informational [Page 17] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +5.7.2. Route Adaptation + + Communication routes MUST be adaptive and converge toward optimality + of the chosen metric (e.g., signal quality, hop count) in time. + +5.7.3. Route Redundancy + + The routing layer SHOULD be configurable to allow secondary and + tertiary routes to be established and used upon failure of the + primary route. + +5.7.4. Route Discovery Time + + Mission-critical commercial applications (e.g., fire, security) + require reliable communication and guaranteed end-to-end delivery of + all messages in a timely fashion. Application-layer time-outs must + be selected judiciously to cover anomalous conditions such as lost + packets and/or route discoveries, yet not be set too large to over- + damp the network response. If route discovery occurs during packet + transmission time (reactive routing), it SHOULD NOT add more than 120 + ms of latency to the packet delivery time. + +5.7.5. Route Preference + + The routing protocol SHOULD allow for the support of manually + configured static preferred routes. + +5.7.6. Real-Time Performance Measures + + A node transmitting a "request with expected reply" to another node + must send the message to the destination and receive the response in + not more than 120 ms. This response time should be achievable with 5 + or less hops in each direction. This requirement assumes network + quiescence and a negligible turnaround time at the destination node. + +5.7.7. Prioritized Routing + + Network and application packet routing prioritization must be + supported to assure that mission-critical applications (e.g., fire + detection) cannot be deferred while less critical applications access + the network. The routing protocol MUST be able to provide routes + with different characteristics, also referred to as Quality of + Service (QoS) routing. + + + + + + + + +Martocci, et al. Informational [Page 18] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +5.8. Security Requirements + + This section sets forth specific requirements that are placed on any + protocols developed or used in the ROLL building environment, in + order to ensure adequate security and retain suitable flexibility of + use and function of the protocol. + + Due to the variety of buildings and tenants, the BMSs must be + completely configurable on-site. + + Due to the quantity of the BMS devices (thousands) and their + inaccessibility (oftentimes above ceilings), security configuration + over the network is preferred over local configuration. + + Wireless encryption and device authentication security policies need + to be considered in commercial buildings, while keeping in mind the + impact on the limited processing capabilities and additional latency + incurred on the sensors, actuators, and controllers. + + BMSs are typically highly configurable in the field, and hence the + security policy is most often dictated by the type of building to + which the BMS is being installed. Single-tenant owner-occupied + office buildings installing lighting or HVAC control are candidates + for implementing a low level of security on the LLN, especially when + the LLN is not connected to an external network. Antithetically, + military or pharmaceutical facilities require strong security + policies. As noted in the installation procedures described in + Sections 3.3 and 5.2, security policies MUST support dynamic + configuration to allow for a low level of security during the + installation phase (prior to building occupancy, when it may be + appropriate to use only diagnostic levels of security), yet to make + it possible to easily raise the security level network-wide during + the commissioning phase of the system. + +5.8.1. Building Security Use Case + + LLNs for commercial building applications should always implement and + use encrypted packets. However, depending on the state of the LLN, + the security keys may either be: + + 1) a key obtained from a trust center already operable on the LLN; + + 2) a pre-shared static key as defined by the general contractor or + its designee; or + + 3) a well-known default static key. + + + + + +Martocci, et al. Informational [Page 19] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + Unless a node entering the network had previously received its + credentials from the trust center, the entering node will try to + solicit the trust center for the network key. If the trust center is + accessible, the trust center will MAC-authenticate the entering node + and return the security keys. If the trust center is not available, + the entering node will check to determine if it has been given a + network key by an off-band means and use it to access the network. + If no network key has been configured in the device, it will revert + to the default network key and enter the network. If neither of + these keys were valid, the device would signal via a fault LED. + + This approach would allow for independent simplified commissioning, + yet centralized authentication. The building owner or building type + would then dictate when the trust center would be deployed. In many + cases, the trust center need not be deployed until all of the local + room commissioning is complete. Yet, at the province of the owner, + the trust center may be deployed from the onset, thereby trading + installation and commissioning flexibility for tighter security. + +5.8.2. Authentication + + Authentication SHOULD be optional on the LLN. Authentication SHOULD + be fully configurable on-site. Authentication policy and updates + MUST be routable over-the-air. Authentication SHOULD occur upon + joining or rejoining a network. However, once authenticated, devices + SHOULD NOT need to reauthenticate with any other devices in the LLN. + Packets may need authentication at the source and destination nodes; + however, packets routed through intermediate hops should not need + reauthentication at each hop. + + These requirements mean that at least one LLN routing protocol + solution specification MUST include support for authentication. + +5.8.3. Encryption + +5.8.3.1. Encryption Types + + Data encryption of packets MUST be supported by all protocol solution + specifications. Support can be provided by use of a network-wide key + and/or an application key. The network key would apply to all + devices in the LLN. The application key would apply to a subset of + devices in the LLN. + + The network key and application key would be mutually exclusive. The + routing protocol MUST allow routing a packet encrypted with an + application key through forwarding devices without requiring each + node in the route to have the application key. + + + + +Martocci, et al. Informational [Page 20] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +5.8.3.2. Packet Encryption + + The encryption policy MUST support either encryption of the payload + only or of the entire packet. Payload-only encryption would + eliminate the decryption/re-encryption overhead at every hop, + providing more real-time performance. + +5.8.4. Disparate Security Policies + + Due to the limited resources of an LLN, the security policy defined + within the LLN MUST be able to differ from that of the rest of the IP + network within the facility, yet packets MUST still be able to route + to or through the LLN from/to these networks. + +5.8.5. Routing Security Policies to Sleeping Devices + + The routing protocol MUST gracefully handle routing temporal security + updates (e.g., dynamic keys) to sleeping devices on their "awake" + cycle to assure that sleeping devices can readily and efficiently + access the network. + +6. Security Considerations + + The requirements placed on the LLN routing protocol in order to + provide the correct level of security support are presented in + Section 5.8. + + LLNs deployed in a building environment may be entirely isolated from + other networks, attached to normal IP networks within the building + yet physically disjoint from the wider Internet, or connected either + directly or through other IP networks to the Internet. Additionally, + even where no wired connectivity exists outside of the building, the + use of wireless infrastructure within the building means that + physical connectivity to the LLN is possible for an attacker. + + Therefore, it is important that any routing protocol solution + designed to meet the requirements included in this document addresses + the security features requirements described in Section 5.8. + Implementations of these protocols will be required in the protocol + specifications to provide the level of support indicated in Section + 5.8, and will be encouraged to make the support flexibly configurable + to enable an operator to make a judgment of the level of security + that they want to deploy at any time. + + As noted in Section 5.8, use/deployment of the different security + features is intended to be optional. This means that, although the + protocols developed must conform to the requirements specified, the + operator is free to determine the level of risk and the trade-offs + + + +Martocci, et al. Informational [Page 21] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + + against performance. An implementation must not make those choices + on behalf of the operator by avoiding implementing any mandatory-to- + implement security features. + + This informational requirements specification introduces no new + security concerns. + +7. Acknowledgments + + In addition to the authors, JP. Vasseur, David Culler, Ted Humpal, + and Zach Shelby are gratefully acknowledged for their contributions + to this document. + +8. References + +8.1. Normative References + + [RFC2119] Bradner, S., "Key words for use in RFCs to Indicate + Requirement Levels", BCP 14, RFC 2119, March 1997. + +8.2. Informative References + + [ROLL-TERM] Vasseur, JP., "Terminology in Low power And Lossy + Networks", Work in Progress, March 2010. + + + + + + + + + + + + + + + + + + + + + + + + + + + +Martocci, et al. Informational [Page 22] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +Appendix A. Additional Building Requirements + + Appendix A contains additional building requirements that were deemed + out of scope for ROLL, yet provided ancillary substance for the + reader. + +A.1. Additional Commercial Product Requirements + +A.1.1. Wired and Wireless Implementations + + Vendors will likely not develop a separate product line for both + wired and wireless networks. Hence, the solutions set forth must + support both wired and wireless implementations. + +A.1.2. World-Wide Applicability + + Wireless devices must be supportable unlicensed bands. + +A.2. Additional Installation and Commissioning Requirements + +A.2.1. Unavailability of an IP Network + + Product commissioning must be performed by an application engineer + prior to the installation of the IP network (e.g., switches, routers, + DHCP, DNS). + +A.3. Additional Network Requirements + +A.3.1. TCP/UDP + + Connection-based and connectionless services must be supported. + +A.3.2. Interference Mitigation + + The network must automatically detect interference and seamlessly + switch the channel to improve communication. Channel changes, and + the nodes' responses to a given channel change, must occur within 60 + seconds. + + + + + + + + + + + + + +Martocci, et al. Informational [Page 23] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +A.3.3. Packet Reliability + + In building automation, it is required that the network meet the + following minimum criteria: + + <1% MAC-layer errors on all messages, after no more than three + retries; + + <0.1% network-layer errors on all messages, after no more than three + additional retries; + + <0.01% application-layer errors on all messages. + + Therefore, application-layer messages will fail no more than once + every 100,000 messages. + +A.3.4. Merging Commissioned Islands + + Subsystems are commissioned by various vendors at various times + during building construction. These subnetworks must seamlessly + merge into networks and networks must seamlessly merge into + internetworks since the end user wants a holistic view of the system. + +A.3.5. Adjustable Routing Table Sizes + + The routing protocol must allow constrained nodes to hold an + abbreviated set of routes. That is, the protocol should not mandate + that the node routing tables be exhaustive. + +A.3.6. Automatic Gain Control + + For wireless implementations, the device radios should incorporate + automatic transmit power regulation to maximize packet transfer and + minimize network interference, regardless of network size or density. + +A.3.7. Device and Network Integrity + + Commercial-building devices must all be periodically scanned to + assure that each device is viable and can communicate data and alarm + information as needed. Routers should maintain previous packet flow + information temporally to minimize overall network overhead. + +A.4. Additional Performance Requirements + +A.4.1. Data Rate Performance + + An effective data rate of 20 kbit/s is the lowest acceptable + operational data rate on the network. + + + +Martocci, et al. Informational [Page 24] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +A.4.2. Firmware Upgrades + + To support high-speed code downloads, routing should support + transports that provide parallel downloads to targeted devices, yet + guarantee packet delivery. In cases where the spatial position of + the devices requires multiple hops, the algorithm should recurse + through the network until all targeted devices have been serviced. + Devices receiving a download may cease normal operation, but upon + completion of the download must automatically resume normal + operation. + +A.4.3. Route Persistence + + To eliminate high network traffic in power-fail or brown-out + conditions, previously established routes should be remembered and + invoked prior to establishing new routes for those devices re- + entering the network. + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Martocci, et al. Informational [Page 25] + +RFC 5867 Building Automation Routing Requirements in LLNs June 2010 + + +Authors' Addresses + + Jerry Martocci + Johnson Controls Inc. + 507 E. Michigan Street + Milwaukee, WI 53202 + USA + Phone: +1 414 524 4010 + EMail: jerald.p.martocci@jci.com + + + Pieter De Mil + Ghent University - IBCN + G. Crommenlaan 8 bus 201 + Ghent 9050 + Belgium + Phone: +32 9331 4981 + Fax: +32 9331 4899 + EMail: pieter.demil@intec.ugent.be + + + Nicolas Riou + Schneider Electric + Technopole 38TEC T3 + 37 quai Paul Louis Merlin + 38050 Grenoble Cedex 9 + France + Phone: +33 4 76 57 66 15 + EMail: nicolas.riou@fr.schneider-electric.com + + + Wouter Vermeylen + Arts Centre Vooruit + Ghent 9000 + Belgium + EMail: wouter@vooruit.be + + + + + + + + + + + + + + + +Martocci, et al. Informational [Page 26] + |