Thread Links Date Links
Thread Prev Thread Next Thread Index Date Prev Date Next Date Index

Re: [8023-CMSG] Server/NIC analogy



Title:

Gary,

I don't know how to "define the target scope of the CMSG to
include multiple stages of switching, multiple hops, and different
link speeds in the same interconnect (or subnet)" and keep this
within 802.3. This kind of scope is at least 802.1 or even higher
up the stack.

The scope for 802.3 is a single link. We can talk about the
length of that link (either in terms of meters or bits for the
potentially high-latency PHYS). We can talk about the system
ramifications of events propagating from bridge to bridge via
multiple links based on the effects of one protocol vs. another
on each individual link. But we cannot talk about multiple
links and expect to keep the scope within 802.3.

Could you elaborate on your suggested scope and whether
you disagree with my comments above or where you think
this project should take place?

Thanks,
Ben

McAlpine, Gary L wrote:
Ben,

I think we can safely assume that, in the not too distant future,
servers will be capable of consuming 10 Gbs links on real applications.
When these servers are used to construct clusters for parallel
processing applications (database processing for instance), they will be
capable of creating frequent congestion with the cluster interconnect.
The traffic loads will include both bursty traffic and high priority
control traffic.

Prioritization will help the high priority traffic get through an
Ethernet cluster interconnect in a timely manner. The rest of the
traffic (that causing all the congestion) is likely to all be
transferred at a low priority. Dropping some portion of it to relieve
congestion is simply unacceptable. During times of heavy loading,
congestion will happen often enough to bring the cluster to its knees
doing retransmissions.

IMO static rate control is also unacceptable because the loading
requirements are not that predictable and not that managable. I just
can't see the database industry buying into static rate control
managment just so they can use Ethernet in backplanes, clusters, and
SANs.

I think the key to making Ethernet acceptable as a short range
interconnect for backplanes and clusters is to support dynamic rate
control. The use of a good implementation of 802.3x (within a limited
enough scope) would be preferrable to drops. If we define the target
scope of the CMSG to include multiple stages of switching, multiple
hops, and different link speeds in the same interconnect (or subnet),
then 802.3x just isn't going to hack it (and neither is static rate
control or packet dropping). I think we need to be thinking in terms of
an improved dynamic rate control, which requires feedback.

I think defining the scope of interconnects we want to target with CM
solutions will go a long way toward framing the range of acceptable
solutions.

Gary

-----Original Message-----
From: owner-stds-802-3-cm@listserv.ieee.org
[mailto:owner-stds-802-3-cm@listserv.ieee.org] On Behalf Of Benjamin
Brown
Sent: Friday, June 04, 2004 2:20 PM
To: STDS-802-3-CM@listserv.ieee.org
Subject: [8023-CMSG] Server/NIC analogy


All,

During a private discussion this afternoon regarding the results of last
week's meeting, the concept of feedback came up - whether it was
necessary or not. There was some level of discussion about this during
the meeting but no one seemed to be able to provide an adequate
justification for providing congestion feedback and why the more common
approach of packet drop wasn't adequate.

During this afternoon's discussion, I came up with something that I
think might be justification. I'm probably just painting a big target on
my chest but let's see how this goes.

Consider a stand-alone server with a 1G Ethernet NIC. Today's CPUs could
easily generate enough traffic to swamp the 1G Ethernet link (okay this
is a bit of an assumption on my part but if they can't today they will
be able to tomorrow). I don't build these things, nor have I looked at
their architecture all that closely in a number of years, but I'll step
out on a limb and state that there's a (most likely proprietary)
mechanism for the NIC to tell the CPU that the link is too slow to
handle all the packets that it is trying to transmit. I'll step even
farther out on that same limb and state that the mechanism is not packet
drop.

Now, let's use this analogy to consider a server card in a backplane
that communicates to the world via a port interface line card. The
server card communicates to the port interface line card using a link
compliant with the newly emerging Backplane Ethernet standard. (Okay, so
I'm looking a little into the future.) If you consider the entire
chassis analogous to the server/NIC in my initial example then it would
seem plausible that you would want to communicate buffer congestion on
the port interface line card back to the server card using a mechanism
other than packet drop.

I'll just close my eyes now. Fire at will.

Ben

--
-----------------------------------------
Benjamin Brown
178 Bear Hill Road
Chichester, NH 03258
603-491-0296 - Cell
603-798-4115 - Office
benjamin-dot-brown-at-ieee-dot-org
(Will this cut down on my spam???)
-----------------------------------------

  

--
-----------------------------------------
Benjamin Brown
178 Bear Hill Road
Chichester, NH 03258
603-491-0296 - Cell
603-798-4115 - Office
benjamin-dot-brown-at-ieee-dot-org
(Will this cut down on my spam???)
-----------------------------------------