Network Working Group S. Bradner Request for Comments: 1944 Harvard University Category: Informational J. McQuaid Bay Networks May 1996 Benchmarking Methodology for Network Interconnect Devices Status of This Memo This memo provides information for the Internet community. This memo does not specify an Internet standard of any kind. Distribution of this memo is unlimited. Abstract This document discusses and defines a number of tests that may be used to describe the performance characteristics of a network interconnecting device. In addition to defining the tests this document also describes specific formats for reporting the results of the tests. Appendix A lists the tests and conditions that we believe should be included for specific cases and gives additional information about testing practices. Appendix B is a reference listing of maximum frame rates to be used with specific frame sizes on various media and Appendix C gives some examples of frame formats to be used in testing. 1. Introduction Vendors often engage in "specsmanship" in an attempt to give their products a better position in the marketplace. This often involves "smoke & mirrors" to confuse the potential users of the products. This document defines a specific set of tests that vendors can use to measure and report the performance characteristics of network devices. The results of these tests will provide the user comparable data from different vendors with which to evaluate these devices. A previous document, "Benchmarking Terminology for Network Interconnect Devices" (RFC 1242), defined many of the terms that are used in this document. The terminology document should be consulted before attempting to make use of this document. Bradner & McQuaid Informational [Page 1] RFC 1944 Benchmarking Methodology May 1996 2. Real world In producing this document the authors attempted to keep in mind the requirement that apparatus to perform the described tests must actually be built. We do not know of "off the shelf" equipment available to implement all of the tests but it is our opinion that such equipment can be constructed. 3. Tests to be run There are a number of tests described in this document. Not all of the tests apply to all types of devices under test (DUTs). Vendors should perform all of the tests that can be supported by a specific type of product. The authors understand that it will take a considerable period of time to perform all of the recommended tests nder all of the recommended conditions. We believe that the results are worth the effort. Appendix A lists some of the tests and conditions that we believe should be included for specific cases. 4. Evaluating the results Performing all of the recommended tests will result in a great deal of data. Much of this data will not apply to the evaluation of the devices under each circumstance. For example, the rate at which a router forwards IPX frames will be of little use in selecting a router for an environment that does not (and will not) support that protocol. Evaluating even that data which is relevant to a particular network installation will require experience which may not be readily available. Furthermore, selection of the tests to be run and evaluation of the test data must be done with an understanding of generally accepted testing practices regarding repeatability, variance and statistical significance of small numbers of trials. 5. Requirements In this document, the words that are used to define the significance of each particular requirement are capitalized. These words are: * "MUST" This word, or the words "REQUIRED" and "SHALL" mean that the item is an absolute requirement of the specification. * "SHOULD" This word or the adjective "RECOMMENDED" means that there may exist valid reasons in particular circumstances to ignore this item, but the full implications should be understood and the case carefully weighed before choosing a different course. * "MAY" This word or the adjective "OPTIONAL" means that this item is truly optional. One vendor may choose to include the item because Bradner & McQuaid Informational [Page 2] RFC 1944 Benchmarking Methodology May 1996 a particular marketplace requires it or because it enhances the product, for example; another vendor may omit the same item. An implementation is not compliant if it fails to satisfy one or more of the MUST requirements for the protocols it implements. An implementation that satisfies all the MUST and all the SHOULD requirements for its protocols is said to be "unconditionally compliant"; one that satisfies all the MUST requirements but not all the SHOULD requirements for its protocols is said to be "conditionally compliant". 6. Test set up The ideal way to implement this series of tests is to use a tester with both transmitting and receiving ports. Connections are made from the sending ports of the tester to the receiving ports of the DUT and from the sending ports of the DUT back to the tester. (see Figure 1) Since the tester both sends the test traffic and receives it back, after the traffic has been forwarded but the DUT, the tester can easily determine if all of the transmitted packets were received and verify that the correct packets were received. The same functionality can be obtained with separate transmitting and receiving devices (see Figure 2) but unless they are remotely controlled by some computer in a way that simulates the single tester, the labor required to accurately perform some of the tests (particularly the throughput test) can be prohibitive. +------------+ | | +------------| tester |<-------------+ | | | | | +------------+ | | | | +------------+ | | | | | +----------->| DUT |--------------+ | | +------------+ Figure 1 +--------+ +------------+ +----------+ | | | | | | | sender |-------->| DUT |--------->| receiver | | | | | | | +--------+ +------------+ +----------+ Figure 2 Bradner & McQuaid Informational [Page 3] RFC 1944 Benchmarking Methodology May 1996 6.1 Test set up for multiple media types Two different setups could be used to test a DUT which is used in real-world networks to connect networks of differing media type, local Ethernet to a backbone FDDI ring for example. The tester could support both media types in which case the set up shown in Figure 1 would be used. Two identical DUTs are used in the other test set up. (see Figure 3) In many cases this set up may more accurately simulate the real world. For example, connecting two LANs together with a WAN link or high speed backbone. This set up would not be as good at simulating a system where clients on a Ethernet LAN were interacting with a server on an FDDI backbone. +-----------+ | | +---------------------| tester |<---------------------+ | | | | | +-----------+ | | | | +----------+ +----------+ | | | | | | | +------->| DUT 1 |-------------->| DUT 2 |---------+ | | | | +----------+ +----------+ Figure 3 7. DUT set up Before starting to perform the tests, the DUT to be tested MUST be configured following the instructions provided to the user. Specifically, it is expected that all of the supported protocols will be configured and enabled during this set up (See Appendix A). It is expected that all of the tests will be run without changing the configuration or setup of the DUT in any way other than that required to do the specific test. For example, it is not acceptable to change the size of frame handling buffers between tests of frame handling rates or to disable all but one transport protocol when testing the throughput of that protocol. It is necessary to modify the configuration when starting a test to determine the effect of filters on throughput, but the only change MUST be to enable the specific filter. The DUT set up SHOULD include the normally recommended routing update intervals and keep alive frequency. The specific version of the software and the exact DUT configuration, including what functions are disabled, used during the tests MUST be included as part of the report of the results. Bradner & McQuaid Informational [Page 4] RFC 1944 Benchmarking Methodology May 1996 8. Frame formats The formats of the test frames to use for TCP/IP over Ethernet are shown in Appendix C: Test Frame Formats. These exact frame formats SHOULD be used in the tests described in this document for this protocol/media combination and that these frames will be used as a template for testing other protocol/media combinations. The specific formats that are used to define the test frames for a particular test series MUST be included in the report of the results. 9. Frame sizes All of the described tests SHOULD be performed at a number of frame sizes. Specifically, the sizes SHOULD include the maximum and minimum legitimate sizes for the protocol under test on the media under test and enough sizes in between to be able to get a full characterization of the DUT performance. Except where noted, at least five frame sizes SHOULD be tested for each test condition. Theoretically the minimum size UDP Echo request frame would consist of an IP header (minimum length 20 octets), a UDP header (8 octets) and whatever MAC level header is required by the media in use. The theoretical maximum frame size is determined by the size of the length field in the IP header. In almost all cases the actual maximum and minimum sizes are determined by the limitations of the media. In theory it would be ideal to distribute the frame sizes in a way that would evenly distribute the theoretical frame rates. These recommendations incorporate this theory but specify frame sizes which are easy to understand and remember. In addition, many of the same frame sizes are specified on each of the media types to allow for easy performance comparisons. Note: The inclusion of an unrealistically small frame size on some of the media types (i.e. with little or no space for data) is to help characterize the per-frame processing overhead of the DUT. 9.1 Frame sizes to be used on Ethernet 64, 128, 256, 512, 1024, 1280, 1518 These sizes include the maximum and minimum frame sizes permitted by the Ethernet standard and a selection of sizes between these extremes with a finer granularity for the smaller frame sizes and higher frame rates. Bradner & McQuaid Informational [Page 5] RFC 1944 Benchmarking Methodology May 1996 9.2 Frame sizes to be used on 4Mb and 16Mb token ring 54, 64, 128, 256, 1024, 1518, 2048, 4472 The frame size recommendations for token ring assume that there is no RIF field in the frames of routed protocols. A RIF field would be present in any direct source route bridge performance test. The minimum size frame for UDP on token ring is 54 octets. The maximum size of 4472 octets is recommended for 16Mb token ring instead of the theoretical size of 17.9Kb because of the size limitations imposed by many token ring interfaces. The reminder of the sizes are selected to permit direct comparisons with other types of media. An IP (i.e. not UDP) frame may be used in addition if a higher data rate is desired, in which case the minimum frame size is 46 octets. 9.3 Frame sizes to be used on FDDI 54, 64, 128, 256, 1024, 1518, 2048, 4472 The minimum size frame for UDP on FDDI is 53 octets, the minimum size of 54 is recommended to allow direct comparison to token ring performance. The maximum size of 4472 is recommended instead of the theoretical maximum size of 4500 octets to permit the same type of comparison. An IP (i.e. not UDP) frame may be used in addition if a higher data rate is desired, in which case the minimum frame size is 45 octets. 9.4 Frame sizes in the presence of disparate MTUs When the interconnect DUT supports connecting links with disparate MTUs, the frame sizes for the link with the *larger* MTU SHOULD be used, up to the limit of the protocol being tested. If the interconnect DUT does not support the fragmenting of frames in the presence of MTU mismatch, the forwarding rate for that frame size shall be reported as zero. For example, the test of IP forwarding with a bridge or router that joins FDDI and Ethernet should use the frame sizes of FDDI when going from the FDDI to the Ethernet link. If the bridge does not support IP fragmentation, the forwarding rate for those frames too large for Ethernet should be reported as zero. 10. Verifying received frames The test equipment SHOULD discard any frames received during a test run that are not actual forwarded test frames. For example, keep- alive and routing update frames SHOULD NOT be included in the count Bradner & McQuaid Informational [Page 6] RFC 1944 Benchmarking Methodology May 1996 of received frames. In any case, the test equipment SHOULD verify the length of the received frames and check that they match the expected length. Preferably, the test equipment SHOULD include sequence numbers in the transmitted frames and check for these numbers on the received frames. If this is done, the reported results SHOULD include in addition to the number of frames dropped, the number of frames that were received out of order, the number of duplicate frames received and the number of gaps in the received frame numbering sequence. This functionality is required for some of the described tests. 11. Modifiers It might be useful to know the DUT performance under a number of conditions; some of these conditions are noted below. The reported results SHOULD include as many of these conditions as the test equipment is able to generate. The suite of tests SHOULD be first run without any modifying conditions and then repeated under each of the conditions separately. To preserve the ability to compare the results of these tests any frames that are required to generate the modifying conditions (management queries for example) will be included in the same data stream as the normal test frames in place of one of the test frames and not be supplied to the DUT on a separate network port. 11.1 Broadcast frames In most router designs special processing is required when frames addressed to the hardware broadcast address are received. In bridges (or in bridge mode on routers) these broadcast frames must be flooded to a number of ports. The stream of test frames SHOULD be augmented with 1% frames addressed to the hardware broadcast address. The frames sent to the broadcast address should be of a type that the router will not need to process. The aim of this test is to determine if there is any effect on the forwarding rate of the other data in the stream. The specific frames that should be used are included in the test frame format document. The broadcast frames SHOULD be evenly distributed throughout the data stream, for example, every 100th frame. The same test SHOULD be performed on bridge-like DUTs but in this case the broadcast packets will be processed and flooded to all outputs. It is understood that a level of broadcast frames of 1% is much higher than many networks experience but, as in drug toxicity evaluations, the higher level is required to be able to gage the Bradner & McQuaid Informational [Page 7] RFC 1944 Benchmarking Methodology May 1996 effect which would otherwise often fall within the normal variability of the system performance. Due to design factors some test equipment will not be able to generate a level of alternate frames this low. In these cases the percentage SHOULD be as small as the equipment can provide and that the actual level be described in the report of the test results. 11.2 Management frames Most data networks now make use of management protocols such as SNMP. In many environments there can be a number of management stations sending queries to the same DUT at the same time. The stream of test frames SHOULD be augmented with one management query as the first frame sent each second during the duration of the trial. The result of the query must fit into one response frame. The response frame SHOULD be verified by the test equipment. One example of the specific query frame that should be used is shown in Appendix C. 11.3 Routing update frames The processing of dynamic routing protocol updates could have a significant impact on the ability of a router to forward data frames. The stream of test frames SHOULD be augmented with one routing update frame transmitted as the first frame transmitted during the trial. Routing update frames SHOULD be sent at the rate specified in Appendix C for the specific routing protocol being used in the test. Two routing update frames are defined in Appendix C for the TCP/IP over Ethernet example. The routing frames are designed to change the routing to a number of networks that are not involved in the forwarding of the test data. The first frame sets the routing table state to "A", the second one changes the state to "B". The frames MUST be alternated during the trial. The test SHOULD verify that the routing update was processed by the DUT. 11.4 Filters Filters are added to routers and bridges to selectively inhibit the forwarding of frames that would normally be forwarded. This is usually done to implement security controls on the data that is accepted between one area and another. Different products have different capabilities to implement filters. Bradner & McQuaid Informational [Page 8] RFC 1944 Benchmarking Methodology May 1996 The DUT SHOULD be first configured to add one filter condition and the tests performed. This filter SHOULD permit the forwarding of the test data stream. In routers this filter SHOULD be of the form: forward input_protocol_address to output_protocol_address In bridges the filter SHOULD be of the form: forward destination_hardware_address The DUT SHOULD be then reconfigured to implement a total of 25 filters. The first 24 of these filters SHOULD be of the form: block input_protocol_address to output_protocol_address The 24 input and output protocol addresses SHOULD not be any that are represented in the test data stream. The last filter SHOULD permit the forwarding of the test data stream. By "first" and "last" we mean to ensure that in the second case, 25 conditions must be checked before the data frames will match the conditions that permit the forwarding of the frame. Of course, if the DUT reorders the filters or does not use a linear scan of the filter rules the effect of the sequence in which the filters are input is properly lost. The exact filters configuration command lines used SHOULD be included with the report of the results. 11.4.1 Filter Addresses Two sets of filter addresses are required, one for the single filter case and one for the 25 filter case. The single filter case should permit traffic from IP address 198.18.1.2 to IP address 198.19.65.2 and deny all other traffic. The 25 filter case should follow the following sequence. deny aa.ba.1.1 to aa.ba.100.1 deny aa.ba.2.2 to aa.ba.101.2 deny aa.ba.3.3 to aa.ba.103.3 ... deny aa.ba.12.12 to aa.ba.112.12 allow aa.bc.1.2 to aa.bc.65.1 deny aa.ba.13.13 to aa.ba.113.13 deny aa.ba.14.14 to aa.ba.114.14 Bradner & McQuaid Informational [Page 9] RFC 1944 Benchmarking Methodology May 1996 ... deny aa.ba.24.24 to aa.ba.124.24 deny all else All previous filter conditions should be cleared from the router before this sequence is entered. The sequence is selected to test to see if the router sorts the filter conditions or accepts them in the order that they were entered. Both of these procedures will result in a greater impact on performance than will some form of hash coding. 12. Protocol addresses It is easier to implement these tests using a single logical stream of data, with one source protocol address and one destination protocol address, and for some conditions like the filters described above, a practical requirement. Networks in the real world are not limited to single streams of data. The test suite SHOULD be first run with a single protocol (or hardware for bridge tests) source and destination address pair. The tests SHOULD then be repeated with using a random destination address. While testing routers the addresses SHOULD be random and uniformly distributed over a range of 256 networks and random and uniformly distributed over the full MAC range for bridges. The specific address ranges to use for IP are shown in Appendix C. 13. Route Set Up It is not reasonable that all of the routing information necessary to forward the test stream, especially in the multiple address case, will be manually set up. At the start of each trial a routing update MUST be sent to the DUT. This routing update MUST include all of the network addresses that will be required for the trial. All of the addresses SHOULD resolve to the same "next-hop". Normally this will be the address of the receiving side of the test equipment. This routing update will have to be repeated at the interval required by the routing protocol being used. An example of the format and repetition interval of the update frames is given in Appendix C. 14. Bidirectional traffic Normal network activity is not all in a single direction. To test the bidirectional performance of a DUT, the test series SHOULD be run with the same data rate being offered from each direction. The sum of the data rates should not exceed the theoretical limit for the media. Bradner & McQuaid Informational [Page 10] RFC 1944 Benchmarking Methodology May 1996 15. Single stream path The full suite of tests SHOULD be run along with whatever modifier conditions that are relevant using a single input and output network port on the DUT. If the internal design of the DUT has multiple distinct pathways, for example, multiple interface cards each with multiple network ports, then all possible types of pathways SHOULD be tested separately. 16. Multi-port Many current router and bridge products provide many network ports in the same module. In performing these tests first half of the ports are designated as "input ports" and half are designated as "output ports". These ports SHOULD be evenly distributed across the DUT architecture. For example if a DUT has two interface cards each of which has four ports, two ports on each interface card are designated as input and two are designated as output. The specified tests are run using the same data rate being offered to each of the input ports. The addresses in the input data streams SHOULD be set so that a frame will be directed to each of the output ports in sequence so that all "output" ports will get an even distribution of packets from this input. The same configuration MAY be used to perform a bidirectional multi-stream test. In this case all of the ports are considered both input and output ports and each data stream MUST consist of frames addressed to all of the other ports. Consider the following 6 port DUT: -------------- ---------| in A out X|-------- ---------| in B out Y|-------- ---------| in C out Z|-------- -------------- The addressing of the data streams for each of the inputs SHOULD be: stream sent to input A: packet to out X, packet to out Y, packet to out Z stream sent to input B: packet to out X, packet to out Y, packet to out Z stream sent to input C packet to out X, packet to out Y, packet to out Z Note that these streams each follow the same sequence so that 3 packets will arrive at output X at the same time, then 3 packets at Y, then 3 packets at Z. This procedure ensures that, as in the real world, the DUT will have to deal with multiple packets addressed to Bradner & McQuaid Informational [Page 11] RFC 1944 Benchmarking Methodology May 1996 the same output at the same time. 17. Multiple protocols This document does not address the issue of testing the effects of a mixed protocol environment other than to suggest that if such tests are wanted then frames SHOULD be distributed between all of the test protocols. The distribution MAY approximate the conditions on the network in which the DUT would be used. 18. Multiple frame sizes This document does not address the issue of testing the effects of a mixed frame size environment other than to suggest that if such tests are wanted then frames SHOULD be distributed between all of the listed sizes for the protocol under test. The distribution MAY approximate the conditions on the network in which the DUT would be used. The authors do not have any idea how the results of such a test would be interpreted other than to directly compare multiple DUTs in some very specific simulated network. 19. Testing performance beyond a single DUT. In the performance testing of a single DUT, the paradigm can be described as applying some input to a DUT and monitoring the output. The results of which can be used to form a basis of characterization of that device under those test conditions. This model is useful when the test input and output are homogenous (e.g., 64-byte IP, 802.3 frames into the DUT; 64 byte IP, 802.3 frames out), or the method of test can distinguish between dissimilar input/output. (E.g., 1518 byte IP, 802.3 frames in; 576 byte, fragmented IP, X.25 frames out.) By extending the single DUT test model, reasonable benchmarks regarding multiple DUTs or heterogeneous environments may be collected. In this extension, the single DUT is replaced by a system of interconnected network DUTs. This test methodology would support the benchmarking of a variety of device/media/service/protocol combinations. For example, a configuration for a LAN-to-WAN-to-LAN test might be: (1) 802.3-> DUT 1 -> X.25 @ 64kbps -> DUT 2 -> 802.3 Or a mixed LAN configuration might be: (2) 802.3 -> DUT 1 -> FDDI -> DUT 2 -> FDDI -> DUT 3 -> 802.3 Bradner & McQuaid Informational [Page 12] RFC 1944 Benchmarking Methodology May 1996 In both examples 1 and 2, end-to-end benchmarks of each system could be empirically ascertained. Other behavior may be characterized through the use of intermediate devices. In example 2, the configuration may be used to give an indication of the FDDI to FDDI capability exhibited by DUT 2. Because multiple DUTs are treated as a single system, there are limitations to this methodology. For instance, this methodology may yield an aggregate benchmark for a tested system. That benchmark alone, however, may not necessarily reflect asymmetries in behavior between the DUTs, latencies introduce by other apparatus (e.g., CSUs/DSUs, switches), etc. Further, care must be used when comparing benchmarks of different systems by ensuring that the DUTs' features/configuration of the tested systems have the appropriate common denominators to allow comparison. 20. Maximum frame rate The maximum frame rates that should be used when testing LAN connections SHOULD be the listed theoretical maximum rate for the frame size on the media. The maximum frame rate that should be used when testing WAN connections SHOULD be greater than the listed theoretical maximum rate for the frame size on that speed connection. The higher rate for WAN tests is to compensate for the fact that some vendors employ various forms of header compression. A list of maximum frame rates for LAN connections is included in Appendix B. 21. Bursty traffic It is convenient to measure the DUT performance under steady state load but this is an unrealistic way to gauge the functioning of a DUT since actual network traffic normally consists of bursts of frames. Some of the tests described below SHOULD be performed with both steady state traffic and with traffic consisting of repeated bursts of frames. The frames within a burst are transmitted with the minimum legitimate inter-frame gap. The objective of the test is to determine the minimum interval between bursts which the DUT can process with no frame loss. During each test the number of frames in each burst is held constant and the inter-burst interval varied. Tests SHOULD be run with burst sizes of 16, 64, 256 and 1024 frames. Bradner & McQuaid Informational [Page 13] RFC 1944 Benchmarking Methodology May 1996 22. Frames per token Although it is possible to configure some token ring and FDDI interfaces to transmit more than one frame each time that the token is received, most of the network devices currently available transmit only one frame per token. These tests SHOULD first be performed while transmitting only one frame per token. Some current high-performance workstation servers do transmit more than one frame per token on FDDI to maximize throughput. Since this may be a common feature in future workstations and servers, interconnect devices with FDDI interfaces SHOULD be tested with 1, 4, 8, and 16 frames per token. The reported frame rate SHOULD be the average rate of frame transmission over the total trial period. 23. Trial description A particular test consists of multiple trials. Each trial returns one piece of information, for example the loss rate at a particular input frame rate. Each trial consists of a number of phases: a) If the DUT is a router, send the routing update to the "input" port and pause two seconds to be sure that the routing has settled. b) Send the "learning frames" to the "output" port and wait 2 seconds to be sure that the learning has settled. Bridge learning frames are frames with source addresses that are the same as the destination addresses used by the test frames. Learning frames for other protocols are used to prime the address resolution tables in the DUT. The formats of the learning frame that should be used are shown in the Test Frame Formats document. c) Run the test trial. d) Wait for two seconds for any residual frames to be received. e) Wait for at least five seconds for the DUT to restabilize. 24. Trial duration The aim of these tests is to determine the rate continuously supportable by the DUT. The actual duration of the test trials must be a compromise between this aim and the duration of the benchmarking test suite. The duration of the test portion of each trial SHOULD be at least 60 seconds. The tests that involve some form of "binary search", for example the throughput test, to determine the exact result MAY use a shorter trial duration to minimize the length of the search procedure, but the final determination SHOULD be made with Bradner & McQuaid Informational [Page 14] RFC 1944 Benchmarking Methodology May 1996 full length trials. 25. Address resolution The DUT SHOULD be able to respond to address resolution requests sent by the DUT wherever the protocol requires such a process. 26. Benchmarking tests: Note: The notation "type of data stream" refers to the above modifications to a frame stream with a constant inter-frame gap, for example, the addition of traffic filters to the configuration of the DUT. 26.1 Throughput Objective: To determine the DUT throughput as defined in RFC 1242. Procedure: Send a specific number of frames at a specific rate through the DUT and then count the frames that are transmitted by the DUT. If the count of offered frames is equal to the count of received frames, the rate of the offered stream is raised and the test rerun. If fewer frames are received than were transmitted, the rate of the offered stream is reduced and the test is rerun. The throughput is the fastest rate at which the count of test frames transmitted by the DUT is equal to the number of test frames sent to it by the test equipment. Reporting format: The results of the throughput test SHOULD be reported in the form of a graph. If it is, the x coordinate SHOULD be the frame size, the y coordinate SHOULD be the frame rate. There SHOULD be at least two lines on the graph. There SHOULD be one line showing the theoretical frame rate for the media at the various frame sizes. The second line SHOULD be the plot of the test results. Additional lines MAY be used on the graph to report the results for each type of data stream tested. Text accompanying the graph SHOULD indicate the protocol, data stream format, and type of media used in the tests. We assume that if a single value is desired for advertising purposes the vendor will select the rate for the minimum frame size for the media. If this is done then the figure MUST be expressed in frames per second. The rate MAY also be expressed in bits (or bytes) per second if the vendor so desires. The Bradner & McQuaid Informational [Page 15] RFC 1944 Benchmarking Methodology May 1996 statement of performance MUST include a/ the measured maximum frame rate, b/ the size of the frame used, c/ the theoretical limit of the media for that frame size, and d/ the type of protocol used in the test. Even if a single value is used as part of the advertising copy, the full table of results SHOULD be included in the product data sheet. 26.2 Latency Objective: To determine the latency as defined in RFC 1242. Procedure: First determine the throughput for DUT at each of the listed frame sizes. Send a stream of frames at a particular frame size through the DUT at the determined throughput rate to a specific destination. The stream SHOULD be at least 120 seconds in duration. An identifying tag SHOULD be included in one frame after 60 seconds with the type of tag being implementation dependent. The time at which this frame is fully transmitted is recorded (timestamp A). The receiver logic in the test equipment MUST recognize the tag information in the frame stream and record the time at which the tagged frame was received (timestamp B). The latency is timestamp B minus timestamp A as per the relevant definition frm RFC 1242, namely latency as defined for store and forward devices or latency as defined for bit forwarding devices. The test MUST be repeated at least 20 times with the reported value being the average of the recorded values. This test SHOULD be performed with the test frame addressed to the same destination as the rest of the data stream and also with each of the test frames addressed to a new destination network. Reporting format: The report MUST state which definition of latency (from RFC 1242) was used for this test. The latency results SHOULD be reported in the format of a table with a row for each of the tested frame sizes. There SHOULD be columns for the frame size, the rate at which the latency test was run for that frame size, for the media types tested, and for the resultant latency values for each type of data stream tested. Bradner & McQuaid Informational [Page 16] RFC 1944 Benchmarking Methodology May 1996 26.3 Frame loss rate Objective: To determine the frame loss rate, as defined in RFC 1242, of a DUT throughout the entire range of input data rates and frame sizes. Procedure: Send a specific number of frames at a specific rate through the DUT to be tested and count the frames that are transmitted by the DUT. The frame loss rate at each point is calculated using the following equation: ( ( input_count - output_count ) * 100 ) / input_count The first trial SHOULD be run for the frame rate that corresponds to 100% of the maximum rate for the frame size on the input media. Repeat the procedure for the rate that corresponds to 90% of the maximum rate used and then for 80% of this rate. This sequence SHOULD be continued (at reducing 10% intervals) until there are two successive trials in which no frames are lost. The maximum granularity of the trials MUST be 10% of the maximum rate, a finer granularity is encouraged. Reporting format: The results of the frame loss rate test SHOULD be plotted as a graph. If this is done then the X axis MUST be the input frame rate as a percent of the theoretical rate for the media at the specific frame size. The Y axis MUST be the percent loss at the particular input rate. The left end of the X axis and the bottom of the Y axis MUST be 0 percent; the right end of the X axis and the top of the Y axis MUST be 100 percent. Multiple lines on the graph MAY used to report the frame loss rate for different frame sizes, protocols, and types of data streams. Note: See section 18 for the maximum frame rates that SHOULD be used. 26.4 Back-to-back frames Objective: To characterize the ability of a DUT to process back-to-back frames as defined in RFC 1242. Procedure: Send a burst of frames with minimum inter-frame gaps to the DUT and count the number of frames forwarded by the DUT. If the count of transmitted frames is equal to the number of frames forwarded the length of the burst is increased and the test is rerun. If Bradner & McQuaid Informational [Page 17] RFC 1944 Benchmarking Methodology May 1996 the number of forwarded frames is less than the number transmitted, the length of the burst is reduced and the test is rerun. The back-to-back value is the number of frames in the longest burst that the DUT will handle without the loss of any frames. The trial length MUST be at least 2 seconds and SHOULD be repeated at least 50 times with the average of the recorded values being reported. Reporting format: The back-to-back results SHOULD be reported in the format of a table with a row for each of the tested frame sizes. There SHOULD be columns for the frame size and for the resultant average frame count for each type of data stream tested. The standard deviation for each measurement MAY also be reported. 26.5 System recovery Objective: To characterize the speed at which a DUT recovers from an overload condition. Procedure: First determine the throughput for a DUT at each of the listed frame sizes. Send a stream of frames at a rate 110% of the recorded throughput rate or the maximum rate for the media, whichever is lower, for at least 60 seconds. At Timestamp A reduce the frame rate to 50% of the above rate and record the time of the last frame lost (Timestamp B). The system recovery time is determined by subtracting Timestamp B from Timestamp A. The test SHOULD be repeated a number of times and the average of the recorded values being reported. Reporting format: The system recovery results SHOULD be reported in the format of a table with a row for each of the tested frame sizes. There SHOULD be columns for the frame size, the frame rate used as the throughput rate for each type of data stream tested, and for the measured recovery time for each type of data stream tested. 26.6 Reset Objective: To characterize the speed at which a DUT recovers from a device or software reset. Bradner & McQuaid Informational [Page 18] RFC 1944 Benchmarking Methodology May 1996 Procedure: First determine the throughput for the DUT for the minimum frame size on the media used in the testing. Send a continuous stream of frames at the determined throughput rate for the minimum sized frames. Cause a reset in the DUT. Monitor the output until frames begin to be forwarded and record the time that the last frame (Timestamp A) of the initial stream and the first frame of the new stream (Timestamp B) are received. A power interruption reset test is performed as above except that the power to the DUT should be interrupted for 10 seconds in place of causing a reset. This test SHOULD only be run using frames addressed to networks directly connected to the DUT so that there is no requirement to delay until a routing update is received. The reset value is obtained by subtracting Timestamp A from Timestamp B. Hardware and software resets, as well as a power interruption SHOULD be tested. Reporting format: The reset value SHOULD be reported in a simple set of statements, one for each reset type. 27. Security Considerations Security issues are not addressed in this document. Bradner & McQuaid Informational [Page 19] RFC 1944 Benchmarking Methodology May 1996 28. Editors' Addresses Scott Bradner Harvard University 1350 Mass. Ave, room 813 Cambridge, MA 02138 Phone +1 617 495-3864 Fax +1 617 496-8500 EMail: sob@harvard.edu Jim McQuaid Bay Networks 3 Federal Street Billerica, MA 01821 Phone +1 508 436-3915 Fax: +1 508 670-8145 EMail: jmcquaid@baynetworks.com Bradner & McQuaid Informational [Page 20] RFC 1944 Benchmarking Methodology May 1996 Appendix A: Testing Considerations A.1 Scope Of This Appendix This appendix discusses certain issues in the benchmarking methodology where experience or judgment may play a role in the tests selected to be run or in the approach to constructing the test with a particular DUT. As such, this appendix MUST not be read as an amendment to the methodology described in the body of this document but as a guide to testing practice. 1. Typical testing practice has been to enable all protocols to be tested and conduct all testing with no further configuration of protocols, even though a given set of trials may exercise only one protocol at a time. This minimizes the opportunities to "tune" a DUT for a single protocol. 2. The least common denominator of the available filter functions should be used to ensure that there is a basis for comparison between vendors. Because of product differences, those conducting and evaluating tests must make a judgment about this issue. 3. Architectural considerations may need to be considered. For example, first perform the tests with the stream going between ports on the same interface card and the repeat the tests with the stream going into a port on one interface card and out of a port on a second interface card. There will almost always be a best case and worst case configuration for a given DUT architecture. 4. Testing done using traffic streams consisting of mixed protocols has not shown much difference between testing with individual protocols. That is, if protocol A testing and protocol B testing give two different performance results, mixed protocol testing appears to give a result which is the average of the two. 5. Wide Area Network (WAN) performance may be tested by setting up two identical devices connected by the appropriate short- haul versions of the WAN modems. Performance is then measured between a LAN interface on one DUT to a LAN interface on the other DUT. The maximum frame rate to be used for LAN-WAN-LAN configurations is a judgment that can be based on known characteristics of the overall system including compression effects, fragmentation, and gross link speeds. Practice suggests that the rate should be at least 110% of the slowest link speed. Substantive issues of testing compression itself are beyond the scope of this document. Bradner & McQuaid Informational [Page 21] RFC 1944 Benchmarking Methodology May 1996 Appendix B: Maximum frame rates reference (Provided by Roger Beeman, Cisco Systems) Size Ethernet 16Mb Token Ring FDDI (bytes) (pps) (pps) (pps) 64 14880 24691 152439 128 8445 13793 85616 256 4528 7326 45620 512 2349 3780 23585 768 1586 2547 15903 1024 1197 1921 11996 1280 961 1542 9630 1518 812 1302 8138 Ethernet size Preamble 64 bits Frame 8 x N bits Gap 96 bits 16Mb Token Ring size SD 8 bits AC 8 bits FC 8 bits DA 48 bits SA 48 bits RI 48 bits ( 06 30 00 12 00 30 ) SNAP DSAP 8 bits SSAP 8 bits Control 8 bits Vendor 24 bits Type 16 bits Data 8 x ( N - 18) bits FCS 32 bits ED 8 bits FS 8 bits Tokens or idles between packets are not included FDDI size Preamble 64 bits SD 8 bits FC 8 bits DA 48 bits SA 48 bits SNAP Bradner & McQuaid Informational [Page 22] RFC 1944 Benchmarking Methodology May 1996 DSAP 8 bits SSAP 8 bits Control 8 bits Vendor 24 bits Type 16 bits Data 8 x ( N - 18) bits FCS 32 bits ED 4 bits FS 12 bits Appendix C: Test Frame Formats This appendix defines the frame formats that may be used with these tests. It also includes protocol specific parameters for TCP/IP over Ethernet to be used with the tests as an example. C.1. Introduction The general logic used in the selection of the parameters and the design of the frame formats is explained for each case within the TCP/IP section. The same logic has been used in the other sections. Comments are used in these sections only if there is a protocol specific feature to be explained. Parameters and frame formats for additional protocols can be defined by the reader by using the same logic. C.2. TCP/IP Information The following section deals with the TCP/IP protocol suite. C.2.1 Frame Type. An application level datagram echo request is used for the test data frame in the protocols that support such a function. A datagram protocol is used to minimize the chance that a router might expect a specific session initialization sequence, as might be the case for a reliable stream protocol. A specific defined protocol is used because some routers verify the protocol field and refuse to forward unknown protocols. For TCP/IP a UDP Echo Request is used. C.2.2 Protocol Addresses Two sets of addresses must be defined: first the addresses assigned to the router ports, and second the address that are to be used in the frames themselves and in the routing updates. The network addresses 192.18.0.0 through 192.19.255.255 are have been assigned to the BMWG by the IANA for this purpose. This Bradner & McQuaid Informational [Page 23] RFC 1944 Benchmarking Methodology May 1996 assignment was made to minimize the chance of conflict in case a testing device were to be accidentally connected to part of the Internet. The specific use of the addresses is detailed below. C.2.2.1 Router port protocol addresses Half of the ports on a multi-port router are referred to as "input" ports and the other half as "output" ports even though some of the tests use all ports both as input and output. A contiguous series of IP Class C network addresses from 198.18.1.0 to 198.18.64.0 have been assigned for use on the "input" ports. A second series from 198.19.1.0 to 198.19.64.0 have been assigned for use on the "output" ports. In all cases the router port is node 1 on the appropriate network. For example, a two port DUT would have an IP address of 198.18.1.1 on one port and 198.19.1.1 on the other port. Some of the tests described in the methodology memo make use of an SNMP management connection to the DUT. The management access address for the DUT is assumed to be the first of the "input" ports (198.18.1.1). C.2.2.2 Frame addresses Some of the described tests assume adjacent network routing (the reboot time test for example). The IP address used in the test frame is that of node 2 on the appropriate Class C network. (198.19.1.2 for example) If the test involves non-adjacent network routing the phantom routers are located at node 10 of each of the appropriate Class C networks. A series of Class C network addresses from 198.18.65.0 to 198.18.254.0 has been assigned for use as the networks accessible through the phantom routers on the "input" side of DUT. The series of Class C networks from 198.19.65.0 to 198.19.254.0 have been assigned to be used as the networks visible through the phantom routers on the "output" side of the DUT. C.2.3 Routing Update Frequency The update interval for each routing protocol is may have to be determined by the specifications of the individual protocol. For IP RIP, Cisco IGRP and for OSPF a routing update frame or frames should precede each stream of test frames by 5 seconds. This frequency is sufficient for trial durations of up to 60 seconds. Routing updates must be mixed with the stream of test frames if longer trial periods are selected. The frequency of updates should be taken from the following table. Bradner & McQuaid Informational [Page 24] RFC 1944 Benchmarking Methodology May 1996 IP-RIP 30 sec IGRP 90 sec OSPF 90 sec C.2.4 Frame Formats - detailed discussion C.2.4.1 Learning Frame In most protocols a procedure is used to determine the mapping between the protocol node address and the MAC address. The Address Resolution Protocol (ARP) is used to perform this function in TCP/IP. No such procedure is required in XNS or IPX because the MAC address is used as the protocol node address. In the ideal case the tester would be able to respond to ARP requests from the DUT. In cases where this is not possible an ARP request should be sent to the router's "output" port. This request should be seen as coming from the immediate destination of the test frame stream. (i.e. the phantom router (Figure 2) or the end node if adjacent network routing is being used.) It is assumed that the router will cache the MAC address of the requesting device. The ARP request should be sent 5 seconds before the test frame stream starts in each trial. Trial lengths of longer than 50 seconds may require that the router be configured for an extended ARP timeout. +--------+ +------------+ | | | phantom |------ P LAN A IN A------| DUT |------------| |------ P LAN B | | OUT A | router |------ P LAN C +--------+ +------------+ Figure 2 In the case where full routing is being used C.2.4.2 Routing Update Frame If the test does not involve adjacent net routing the tester must supply proper routing information using a routing update. A single routing update is used before each trial on each "destination" port (see section C.24). This update includes the network addresses that are reachable through a phantom router on the network attached to the port. For a full mesh test, one destination network address is present in the routing update for each of the "input" ports. The test stream on each Bradner & McQuaid Informational [Page 25] RFC 1944 Benchmarking Methodology May 1996 "input" port consists of a repeating sequence of frames, one to each of the "output" ports. C.2.4.3 Management Query Frame The management overhead test uses SNMP to query a set of variables that should be present in all DUTs that support SNMP. The variables for a single interface only are read by an NMS at the appropriate intervals. The list of variables to retrieve follow: sysUpTime ifInOctets ifOutOctets ifInUcastPkts ifOutUcastPkts C.2.4.4 Test Frames The test frame is an UDP Echo Request with enough data to fill out the required frame size. The data should not be all bits off or all bits on since these patters can cause a "bit stuffing" process to be used to maintain clock synchronization on WAN links. This process will result in a longer frame than was intended. C.2.4.5 Frame Formats - TCP/IP on Ethernet Each of the frames below are described for the 1st pair of DUT ports, i.e. "input" port #1 and "output" port #1. Addresses must be changed if the frame is to be used for other ports. C.2.6.1 Learning Frame ARP Request on Ethernet -- DATAGRAM HEADER offset data (hex) description 00 FF FF FF FF FF FF dest MAC address send to broadcast address 06 xx xx xx xx xx xx set to source MAC address 12 08 06 ARP type 14 00 01 hardware type Ethernet = 1 16 08 00 protocol type IP = 800 18 06 hardware address length 48 bits on Ethernet 19 04 protocol address length 4 octets for IP 20 00 01 opcode request = 1 22 xx xx xx xx xx xx source MAC address 28 xx xx xx xx source IP address Bradner & McQuaid Informational [Page 26] RFC 1944 Benchmarking Methodology May 1996 32 FF FF FF FF FF FF requesting DUT's MAC address 38 xx xx xx xx DUT's IP address C.2.6.2 Routing Update Frame -- DATAGRAM HEADER offset data (hex) description 00 FF FF FF FF FF FF dest MAC address is broadcast 06 xx xx xx xx xx xx source hardware address 12 08 00 type -- IP HEADER 14 45 IP version - 4, header length (4 byte units) - 5 15 00 service field 16 00 EE total length 18 00 00 ID 20 40 00 flags (3 bits) 4 (do not fragment), fragment offset-0 22 0A TTL 23 11 protocol - 17 (UDP) 24 C4 8D header checksum 26 xx xx xx xx source IP address 30 xx xx xx destination IP address 33 FF host part = FF for broadcast -- UDP HEADER 34 02 08 source port 208 = RIP 36 02 08 destination port 208 = RIP 38 00 DA UDP message length 40 00 00 UDP checksum -- RIP packet 42 02 command = response 43 01 version = 1 44 00 00 0 -- net 1 46 00 02 family = IP 48 00 00 0 50 xx xx xx net 1 IP address 53 00 net not node 54 00 00 00 00 0 58 00 00 00 00 0 62 00 00 00 07 metric 7 -- net 2 Bradner & McQuaid Informational [Page 27] RFC 1944 Benchmarking Methodology May 1996 66 00 02 family = IP 68 00 00 0 70 xx xx xx net 2 IP address 73 00 net not node 74 00 00 00 00 0 78 00 00 00 00 0 82 00 00 00 07 metric 7 -- net 3 86 00 02 family = IP 88 00 00 0 90 xx xx xx net 3 IP address 93 00 net not node 94 00 00 00 00 0 98 00 00 00 00 0 102 00 00 00 07 metric 7 -- net 4 106 00 02 family = IP 108 00 00 0 110 xx xx xx net 4 IP address 113 00 net not node 114 00 00 00 00 0 118 00 00 00 00 0 122 00 00 00 07 metric 7 -- net 5 126 00 02 family = IP 128 00 00 0 130 00 net 5 IP address 133 00 net not node 134 00 00 00 00 0 138 00 00 00 00 0 142 00 00 00 07 metric 7 -- net 6 146 00 02 family = IP 148 00 00 0 150 xx xx xx net 6 IP address 153 00 net not node 154 00 00 00 00 0 158 00 00 00 00 0 162 00 00 00 07 metric 7 C.2.4.6 Management Query Frame To be defined. Bradner & McQuaid Informational [Page 28] RFC 1944 Benchmarking Methodology May 1996 C.2.6.4 Test Frames UDP echo request on Ethernet -- DATAGRAM HEADER offset data (hex) description 00 xx xx xx xx xx xx set to dest MAC address 06 xx xx xx xx xx xx set to source MAC address 12 08 00 type -- IP HEADER 14 45 IP version - 4 header length 5 4 byte units 15 00 TOS 16 00 2E total length* 18 00 00 ID 20 00 00 flags (3 bits) - 0 fragment offset-0 22 0A TTL 23 11 protocol - 17 (UDP) 24 C4 8D header checksum* 26 xx xx xx xx set to source IP address** 30 xx xx xx xx set to destination IP address** -- UDP HEADER 34 C0 20 source port 36 00 07 destination port 07 = Echo 38 00 1A UDP message length* 40 00 00 UDP checksum -- UDP DATA 42 00 01 02 03 04 05 06 07 some data*** 50 08 09 0A 0B 0C 0D 0E 0F * - change for different length frames ** - change for different logical streams *** - fill remainder of frame with incrementing octets, repeated if required by frame length Bradner & McQuaid Informational [Page 29] RFC 1944 Benchmarking Methodology May 1996 Values to be used in Total Length and UDP message length fields: frame size total length UDP message length 64 00 2E 00 1A 128 00 6E 00 5A 256 00 EE 00 9A 512 01 EE 01 9A 768 02 EE 02 9A 1024 03 EE 03 9A 1280 04 EE 04 9A 1518 05 DC 05 C8 Bradner & McQuaid Informational [Page 30]
RFC, FYI, BCP