|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: Concensus Call on Urgent Pointer.Michael I done the calculations for 100Mbits/sec to 100Gbps for the 6 switch topology you describe and 50KM distance. The data is below. (I know it's a bit of rock fetching but what the hell!) A couple of points to clear up from your previous email. 1/ I had allowed 25uS end system processing delay - you suggested 4 - 5us. Hardware can process a lot faster than even what you suggest but even when the hardware has finished processing an incoming header and wants to send an ACK in reply, the outbound port can be busy sending a maximum frame => delay of 12uS. Secondly - we only send an ACK for every second frame hence another 12uS possible. 2/ Can you give me the source of the error rates you suggest I use. >the packet drop rate due to either corruption (very low rates should be >assumed - at least 10E-9 to 10E-12) or congestion which ranges from 1-3% in >the Internet (depending upon which set of measurements you go by) and will The bit error rate on worst case copper at Gbps is 10E-10 => PACKET drop is 10e-6 i.e. 1 in a million(1500bytes). We've verified similiar (but better) rates in our Lab's using broadcom phi's. 3/ The packet drop rate due to congestion - in a LAN certainly - can be reduced to 0 by turning on flow control. In general flow control can be bad as it's easy to think of scenarios where lockup happens, but in certain instances it can be useful to turn it on. 4/ Vern's numbers w.r.t. corrupt packets on FDDI which were not being caught by CRC is interesting. We've ran our 12 port gigabit switches for days at full line rate with a random scattering of packet sizes and did not see any internal packet loss or corruption. We'ed normally run such data corruption tests in flow control mode to allow us notice if even one packet went missing over days and why it went missing. So in a LAN scenario the maximum loss rate can be configured to be related to the bit error rate. This is also possible for a MAN environment or rather a disaster recovery link at least. 5/Vern - the formula you had in your email is interesting - Can you give me a pointer to its derivation. I would also like to see formulas which had the line rate and window size as parameters. The formula in your email aims to obtain a theoretical max throughput. This is interesting. However on a Gigabit line if the error rate is 0 the throughput is not infinite , it is 1Gbps. Also, on a Gbps line if the packet error rate is 10E-6 and the window is of infinite size then the throughput is not affected much. 6/ From the data below the astute reader will have noticed:- a/ Keeping 10 trunked 100mbit/sec lines saturated requires more memory over same distances using store and forward switching elements than 1 Gbps line. b/ Keeping 10 trunked 1Gbps lines saturated requires more memory than 1x10Gbps line. But the differences are reducing as the delay due to the cable becomes more evident. c/Between 10Gbps and 100Gbps the memory requirements compare. d/Cut through switching was interesting at 10Mbps when the cut-through time was 30-40uS compared to 1.2mS store and forward delay for max size packet. At 100Mbps it is not interesting - 30-40uS (completely empty switch) versus 120uS store and forward type delay. e/I've put the switch delay at 0.5 the packet store and forward delay. For true switches that attempt to handle congestion cases (small bursts of congestion at least) then internal switching must be faster than the line rate. However I've not looked in detail or measured any multiport 10G or 100G switches. 7/ I still don't see the required memory space cost as a big issue for hardware implementations since h/w can process acks fast and keep window small as possible. TOE chips that support iSCSI and OTHER protocols at 1Gbps over TCP will have to have appropriate memory on board in any case for those OTHER protocols. It's cheaper to have one TOE supporting BOTH than TWO TOE adaptoes - one with some fractional smaller amount of memory and another fully fledged TOE adaptor for general purpose TCP connections. 7/My take on seeing the bunch of emails from the list and Randall Stewarts work leaves me with the impression that the debate has shifted from a MUST - MAY debate to a MAY - DELETE debate. I think that should be the question i.e. should this be deleted or left in as a MAY. Dick Gahan 3Com 0.1 0.1 1 1 10 10 100 100 Protocol bandwidth in Gbps 1500 1500 1500 1500 1500 1500 1500 1500 Max frame size in bytes 120 120 12 12 1.2 1.2 0.12 0.12 time to transmit frame in uS 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 Signal speed on cable w.r.t c 700 50000 700 50000 700 50000 700 50000 Total length of cable in METERS 5 333 5 333 5 333 5 333 Total Delay due to cable in uS 7 7 7 7 7 7 7 7 Number of forwarding occurences 840 840 84 84 8.4 8.4 0.84 0.84 Total delay due to forwarding 6 6 6 6 6 6 6 6 Number of switches 60 60 6 6 0.6 0.6 0.06 0.06 Switch latency in uS - half pkt time 360 360 36 36 3.6 3.6 0.36 0.36 total switching latency delay uS 240 240 24 24 2.4 2.4 0.24 0.24 End system TCP delay - min + a bit => 2 pkts worth 1445 1773 149 477 19 348 6 335 Total one way delay 2889 3547 297 955 38 695 12 670 Round Trip delay uS 24 30 25 80 32 580 102 5580 Max Ethernet 1500 byte frames on Wire/Window 36 44 37 119 48 869 153 8369 Bandwidth/Delay product in KB 1E-10 1E-10 1E-12 1E-12 1E-12 1E-12 Bit error rate per link (10G and 100G is fibre only) 1.20E-061.20E-061.20E-081.20E-081.20E-081.20E-08Packet Error rate per link for 1500 bytes 8.40E-068.40E-068.40E-088.40E-088.40E-088.40E-08Packet Error rate end to end due to line loss alone PLANET PROJECT will connect millions of people worldwide through the combined technology of 3Com and the Internet. Find out more and register now at http://www.planetproject.com
Home Last updated: Tue Sep 04 01:06:18 2001 6315 messages in chronological order |