Kim:
Any thoughts?
OK, I'll bite. CAUTION: As always, my email is long, wordy, technical and sometimes skirts off topic; however, I've got to put up with free marketing references to Cisco/Juniper at every turn on NANOG. It's nice to get Foundry's name here every once in awhile. We're all good companies. For the most part (95%), I'll stay on HA operational topics for the NANOG reader. Some recommended books on this subject are listed below. I will refer to these books during this email: Top Down Network Design by Priscilla Oppenheimer (Cisco Press) Designing Enterprise Routing and Switching Architectures by Howard Berkowitz (McMillan Press) High Availability Network Fundamentals by Chris Oggerino (Cisco Press) Lot's of other references to include industry standards by Telcordia (how's your calculus?). See HA Networking Fundamentals for a good root reference list. I guess the main thing to do is look on page 48 of Priscilla's book. She categorizes customer requirements and recommends a method to prioritize those requirements for network design tradeoffs. I've added "profitability" to the list for service providers. You can tailor it to your business goals. I go back to this a lot and it helps me know where availability sits as a design requirement. Now we have to think about what you mean by "Industry Standard". This definitely depends on the industry; however, it varies per company based on design requirements mapped to business goals of YOUR company in that industry. Obviously, having better availability is just one part of a multi-faceted competitive business plan. In some industries, it is assumed to be basic to have high availability. Other's ignore it. Some components you must look at are: Human Error/Process Error Physical Infrastructure Security and Robustness Equipment Quality Technologies Special Events, Risks, and Threats (Sean Donelan digging up your fiber, hacker attack, governmental or organizational shutdown/de-peering, war or political unrest, resource shortages, economy, insert your imagination here) Maintenance On the technology side, basically... The lower you push your redundant failover technology, the better your failover. SONET APS can failover in 50ms over and over again. L2 and L3 protocols continue to operate as normal with minimum In-Flight Packet loss (IFP). This is exactly why the 10GbE Forum is promoting APS in the 10GbE WAN PHY! Foundry Networks (my company) has two new technologies that can give you sub second failover and avoid the failover of L3 and slower L2 redundancy protocols (RSTP and STP). The technologies are Metro Ring Protocol and Virtual Switch Redundancy Protocol. Both of these are currently in beta (soon to be released), but I've been playing with them for the past week. VSRP is VRRP on L2 steroids (sub second failover). Easy to understand (one switch is actively forwarding while the other is on standby). All of these L2 protocols are interoperable on the same devices in the same networks (RSTP, STP, VSRP, MRP). A customer can run STP with a provider VSRP edge and MRP core. VLAN stacking and STP tunneling is supported for those of you looking at Metro business plans. Below is an example of HA technology with MRP. Take a look at this topology: _____P1A PE1 1 | ___P2A___P2C \_____P1B/ 2 | _____P4A \___P2B___P3A/ 3 | \_____PE2 I've got link P2B to P3A running MPLS (LER to LER, don't ask why, it's just a lab) OC-48 (wire speed 2.5G) with Draft Martini L2 VPN. Link P2A to P2C is 802.3ae draft 4.2 compliant 10GbE. All other links are GbE. I've got 50 VLAN's. 25 of them travel clockwise around the rings and 25 of them travel counter clockwise. Each group of 25 is grouped in a topology group and run an instance of MRP on the lead (master) VLAN of that topology group. Rings are 1, 2, and 3. I really hope my diagram shows up OK for the readers of this email. The MRP ring masters are PE1 for ring1, P1B for ring 2, and PE2 for ring 3. MRP masters send out Ring Health Packets (RHP) around the ring every 100ms (configurable). They originate these out of their primary ports and receive them on their secondary ports. MRP masters block forwarding on their secondary ports if they receive the RHP's. They transition to forwarding (ring broken) when they stop receiving the RHP's. Now let's assume that all traffic is taking the bottom path via MRP primary paths on the masters. OK, let's start pinging (192.168.1.40 is PE2 loopback address): PE1#ping 192.168.1.40 count 100000000 time 800 Sending 100000000, 16-byte ICMP Echo to 192.168.1.40, timeout 800 msec, TTL 64 Type Control-c to abort 511000Request timed out. < Here I unplug PE1 to P1B link (primary path). 1 In-Flight Packet (IFP) lost. 854000Request timed out. < Here I unplug PE1B to P2B link (primary path) 1 IFP lost 1160000Request timed out.< Here I unplug P3A to PE2 link (primary path) 2 IFP's lost. All traffic on secondary path now. Request timed out. 1513000Request timed out. < Here I plug PE2 to PE3 link back. 2 IFP's lost. Request timed out. 1638000Request timed out. < Plug P1B to P2B link back in. 1 IFP lost. 1823000Request timed out. < Plug PE1 to P1B link back. 2 IFP's lost. Request timed out. 1^674000 Not too bad considering that MRP is a software technology eh? Also, the CPU's of all the devices are at 1%! 802.17 Resilient Packet Ring (RPR) is supposed to do EXACTLY what MRP does, but faster 'cause it is in HW. Personally, I don't think the industry needs another L2 technology. Ethernet will be just fine with APS in the WAN PHY (Coming this year I'm sure)! RPR is not Ethernet and will be more expensive. I was a Token Ring fan. I've learned to respect Ethernet and I regard Ethernet as the clear winner. My XBOX(TM) at home has Ethernet (NOTE: My XBOX has only rebooted suddenly on me 3 times as opposed to ZERO for my PS2. Thanks MS!)! ATM Segmentation and Reassembly on OC-192 will be a lot more costly than simple 10GbE as well. I'm not even sure if SAR has the capability to do it at wire speed today. I've seen nothing on this from the ATM front. Ethernet will be at 40GbE (OC-768) before ATM SAR is at OC-192. My money is on Ethernet. LINX is just one of many folks running 10GbE today! Took them 3 minutes to make the conversion from what I read in the press release. Wonder how long it would take to do an RPR upgrade from GbE (Haven't seen a working RPR network yet. I have seen MRP on 10GbE, GbE, and POS). ATM? We can see here that the technology can get us to the point of 100% availability (I don't consider one or two packets per user session on a link failover as downtime. Do you?); however, as you can see from my design, I've got Single Points of Failure. I can easily design more rings (at more cost) and remove these SPOF's. Now the only question is: What are your business goals and your acceptable amount of downtime. I want to point you to Howard Berkowitz's book for some advice on downtime tolerance. I don't want to explain it here; however, the unit of measurement is called the "Fulford". Howard talks about a network design requirement no more than two Fulfords a year. Hilarious scenario, but often true. Howard also is quick to point out that simply having redundant links does not equal high availability! Good read (although Howard can get a bit repetitious, it helps drive the main points home). I think that there is no acceptable industry standard that you can simply overlay into an individual company's requirements. It is all customized. Some folks are happy with slow convergence. As long as the phone doesn't ring. Some users accept a provider to be down for 30 minutes every Sunday night. All relevant. Some providers have governmental reporting requirements if they have downtime. One other thing. If an organization doesn't have downtime reporting processes and run charts, then I feel they really don't know what they've got. You can calculate device serial and parallel availability by using MTBF/MTTR calculations. These are all probabilities. There is a much bigger picture. How many times does a network engineer mess something up and then hide it? I think every one on this list has made a serious mistake and caused network downtime WITHOUT reporting it. I caused about 10 seconds of downtime on a large service provider network by accidentally removing the default route (fat fingered!) not two months ago. The guy in charge said: "Only 10 seconds? Don't worry 'bout it. Nobody will every know.". Now you know. Plan, Do, Check, Act, Repeat. You must have processes to track and improve your availability. Else you are doing nothing but talking 'bout it and you are still clueless. Bottom line, there is no hard and fast rule. Everyone wants 5 nines (99.999%), yet how can you get that using routing protocol (L3) or spanning tree (L2) redundancy technologies! I'm a big L3 fan in my designs, but I understand the convergence factors. VSRP on an L2 core could give you sub second. L3 BGP may be your only choice. OSPF with link aggregation and auto-cost decrementation (is this a word?) on link failure (hey, aggregate 4X10GbE and loosing one link can have significant impact on a network core). An example is, you can get 50ms to 5 seconds failover using Rapid Spanning Tree, but it takes the normal failover of STP when returning back to the original path. This is almost a minute of downtime. Not too good on the statistics. I worked on US Military networks before. Their redundancy is not only terrestrial, but uses aircraft and satellites. Wanna buy a used AWACS? See how it all relates? How much money you got? What are your goals? How smart are your humans (training is the most upstream process)? Save money and use monkeys? Now back to your original question:
Is this a hard and fast rule or is this a value that we all try and emulate as best we can? Do I have the value incorrect? Is it higher or lower?
Set your own standard. I doubt if you'll find the right answer on NANOG. If you want my generic answer. I'd say you want 99.999% availability from all network endpoints to network endpoints during times of network utilization. I doubt if you'll hear many complaints from users/customers at this level. Please be careful when jumping this high. You could pull a muscle (take away from another key requirement such as Cost, Manageability, Security, Reliability, et al..). Gary Blankenship Systems Engineer - Japan