WO1999066758A2 - An interconnect network for operation within a communication node - Google Patents

An interconnect network for operation within a communication node Download PDF

Info

Publication number
WO1999066758A2
WO1999066758A2 PCT/US1999/013641 US9913641W WO9966758A2 WO 1999066758 A2 WO1999066758 A2 WO 1999066758A2 US 9913641 W US9913641 W US 9913641W WO 9966758 A2 WO9966758 A2 WO 9966758A2
Authority
WO
WIPO (PCT)
Prior art keywords
local
interconnect
information
channels
modules
Prior art date
Application number
PCT/US1999/013641
Other languages
French (fr)
Other versions
WO1999066758A3 (en
Inventor
Frank Kastenholz
Tom Westberg
Steven R. Willis
Original Assignee
Unisphere Solutions, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Unisphere Solutions, Inc. filed Critical Unisphere Solutions, Inc.
Priority to IL13461199A priority Critical patent/IL134611A/en
Priority to EP99930340A priority patent/EP1066735B1/en
Priority to CA002301910A priority patent/CA2301910A1/en
Priority to AU46896/99A priority patent/AU760640B2/en
Publication of WO1999066758A2 publication Critical patent/WO1999066758A2/en
Publication of WO1999066758A3 publication Critical patent/WO1999066758A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/61Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/74Address processing for routing
    • H04L45/742Route cache; Operation thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q11/00Selecting arrangements for multiplex systems
    • H04Q11/04Selecting arrangements for multiplex systems for time-division multiplexing
    • H04Q11/0428Integrated services digital network, i.e. systems for transmission of different types of digitised signals, e.g. speech, data, telecentral, television signals
    • H04Q11/0478Provisions for broadband connections
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5614User Network Interface
    • H04L2012/5618Bridges, gateways [GW] or interworking units [IWU]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • H04L2012/5646Cell characteristics, e.g. loss, delay, jitter, sequence integrity
    • H04L2012/5652Cell construction, e.g. including header, packetisation, depacketisation, assembly, reassembly
    • H04L2012/5653Cell construction, e.g. including header, packetisation, depacketisation, assembly, reassembly using the ATM adaptation layer [AAL]
    • H04L2012/5658Cell construction, e.g. including header, packetisation, depacketisation, assembly, reassembly using the ATM adaptation layer [AAL] using the AAL5
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • H04L2012/5665Interaction of ATM with other protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • H04L2012/5665Interaction of ATM with other protocols
    • H04L2012/5667IP over ATM
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • H04L2012/5665Interaction of ATM with other protocols
    • H04L2012/5669Multiprotocol over ATM [MPOA]

Definitions

  • the invention relates generally to communication nodes, and more particularly to an interconnect network for operation within a communication node.
  • ISDN Integrated Services Digital Network
  • ATM Asynchronous Transfer Mode
  • IP Internet Protocol
  • ISDN an early attempt at a multi-service architecture (i.e., an architecture capable of handling a variety of communication encapsulations), which is based on the telephone hierarchy, apportions bandwidth in 64 kilobits per second (Kbps) circuits.
  • LANs local area networks
  • Mbps megabits per second
  • ATM is a packet switching protocol that was conceived as a transport mechanism for broadband ISDN. ATM transfers information in fixed-length packets called cells.
  • the cells travel over virtual connections (VCs) between communication nodes that are established prior to each communication session.
  • VCs virtual connections
  • the combination of fixed cell formats and VCs renders ATM a faster alternative to ISDN. Additionally, ATM handles bursts of data traffic more efficiently than time division multiplexing (TDM) and provides high-quality voice and video support.
  • TDM time division multiplexing
  • WWW World Wide Web
  • FR frame relay
  • ISPs Internet Service Providers
  • ISDN provides a relatively low-speed access solution.
  • ATM supports frame relay, virtual private networks (VPNs), circuit emulation, private branch exchange (PBX) interconnects and quality of service (QoS), but does not mesh easily with existing data protocols.
  • IP supports applications such as Internet Access and VPNs, for which cost connectivity is important.
  • IP has yet to demonstrate industrial-strength reliability.
  • full service providers find it necessary to maintain parallel switching networks. Because such parallel networks require maintenance and service of a variety of devices such as, voice switches, frame relay switches, ATM switches, routers, add/drop multiplexers, and digital cross-connects, they have a high associated capital equipment cost.
  • Conventional communication nodes also have a variety of drawbacks. For example, conventional communication nodes fail to provide sufficient ease of scalability. Typically, conventional switching nodes provide a switching/routing network having a fixed bandwidth. However, as enterprises grow, their needs also grow. But, the fixed bandwidth switching routing network of conventional technology requires enterprises to predict such growth and purchase systems having sufficiently large bandwidth up front; thereby compounding the challenge of maintaining parallel networks. Another drawback of conventional systems is reliability. Because conventional systems fail to provide a single switch/routing network that can operate on a variety of protocols, today's Giga Points-of-Presence (GigaPoPs) and Access PoPs are a complex and expensive aggregation of core routers connecting smaller Access PoPs to the core transport capacity. These structures are fragile, with frequent service outages due to performance limitations and equipment failures. Enterprises cannot afford to be exposed to significant down time due to failures or updates associated with conventional technology.
  • GigaPoPs Giga Points-of-Presence
  • the switching/routing networks of conventional systems are typically designed to operate under the constraints of a particular protocol, they lack the flexibility to adapt to emerging technologies, employing new communication protocols. As discussed above, different protocols provide different QoS features. Thus, another drawback of a network operating under the constraints of a single protocol is that a service provider cannot offer varying grades of service to users having differing priority requirements; thus causing service providers to forego a potentially significant source of revenue. Accordingly, it is an object of the invention to provide an interconnect network that enables a multi-service communication node to handle a variety of communication protocols, without requiring the maintenance of costly parallel networks. It is a further object of the invention to provide an interconnect network that enables a communication node to adapt to communication protocols employed by emerging technologies.
  • Another object of the invention is to provide a scalable interconnect network enabling bandwidth scaling of a communication node to fit the needs of providers having varying bandwidth requirements.
  • An additional object of the invention is to provide a fault-tolerant interconnect network capable of repair and update, without causing down-time or compromising operation of the communication node.
  • the invention is directed to communication nodes. More particularly, it is directed to interconnect networks in communication nodes.
  • a communication node includes interconnect networks that enable the node to transfer a variety of communication protocols.
  • an interconnect network according to the invention enables a communication node to handle ATM and IP Packet-over-SONET protocols with the same hardware.
  • An interconnect network according to an additional embodiment of the invention also enables a communication node to provide Frame Relay Data Terminal Equipment (DTE) and Multiprotocol Label Switching(MPLS) functionality.
  • DTE Frame Relay Data Terminal Equipment
  • MPLS Multiprotocol Label Switching
  • An interconnect network according to another embodiment of the invention enables a communication node to act as both a native ATM switch and a native IP router, operating at line speeds up to at least as high as 2.488 Gps (OC48c/STM16c).
  • the invention can provide improved reliability.
  • the invention provides Automatic Protection Switching (APS), wherein Open Systems Interconnection (OSI) Layer 2 and Layer 3 information is mirrored to provide rapid APS switchover.
  • OSI Open Systems Interconnection
  • system modules can be hot-swappable, and designed so that single component failures do not lead to total node failure.
  • the communication node is packaged in a scalable set of modules. OC48 line cards and Gigabit Ethernet modules populate a local communication interface module including a local interconnect network.
  • An optional front end access module provides fan out to OC12/STM4, OC3/STM1, DS3, or E3 interfaces, and an optional expanded interconnect module, sometimes referred to as a hyperconnect fabric, allows dynamic bandwidth expansion of the communication node to include up to eight interconnected local interconnect modules, thereby providing 160 Gbs of essentially non-blocking bandwidth.
  • Yet another embodiment of the invention enables service providers to offer enterprises differing grades or quality of service (QoS).
  • QoS quality of service
  • an interconnect network is incorporated in a communication node having a local communication interface, an associated local interconnect network, and scaling elements.
  • the local communication interface includes a plurality of external communication channels for coupling information into and out of the node and a plurality of internal communication channels for transferring information within the node. Each external communication channel couples to an internal communication channel.
  • the local interconnect network has local transfer elements for directing information between the internal communication channels, and consequently between the external communication channels.
  • the scaling elements enable dynamically scaling the node to include additional local communication interfaces having additional associated local interconnect networks, such that information can be transferred between the local communication interfaces.
  • the communication node optionally, can transfer information between any of the internal communication channels, and thus any of the external communication channels, of the local communication interfaces.
  • the communication node provides an additional local communication interface and an additional local interconnect network; and the scaling elements include an expanded interconnect network.
  • the additional local communication interface has an additional plurality of external communication channels for coupling information in and out of the node, and an additional plurality of internal communication channels for transferring information within the node.
  • the internal and external communication channels of the additional local communication interface couple to each other.
  • the additional local interconnect network includes additional local transfer elements for directing information between the additional plurality of internal communication channels.
  • the local interconnect network and the additional local interconnect network both include non-local transfer elements for directing information between the internal communication channels and the expanded interconnect network.
  • the expanded interconnect network includes expanded transfer elements for directing information between the local interconnect networks, such that information, optionally, can be transferred between any of the internal communication channels of the local communication interfaces.
  • the invention includes up to eight local communication interfaces, with associated local interconnect networks.
  • the expanded interconnect network remains unchanged, regardless of the number of local communication interfaces, and provides the ability to transfer information between the internal communication channels.
  • the communication node can be scaled to change the number of local communication interfaces, while the node is operating transferring information. In this way, a communication node, incorporating an interconnect network according to one embodiment of the invention, can more easily meet a service provider's varying bandwidth needs.
  • an interconnect network can monitor the availability of communication channels. More particularly, the local interconnect network can include a plurality of transceivers for transferring information between the local transfer elements and the internal communication channels. Each transceiver couples to an associated internal communication channel, and has a corresponding availability status indicative of an availability of that communication channel for transferring information.
  • the local interconnect network may also include a plurality of memory storage queues, having associated ones of the transceivers, and including memory for storing information to be transferred by an associated transceiver.
  • the interconnect networks may further include control elements for setting the status corresponding to a particular internal communication channel to indicate unavailability for transferring information, in response to an associated memory queue reaching a selectable content level. In this way, the communication node lowers the likelihood of losing information or blocking transfer due to overloading a particular channel. Additionally, according to a further embodiment, the information coupled into the communication node is assigned a particular priority, and the interconnect networks optionally includes control elements for setting the status corresponding to a particular channel to indicate availability for receiving information having a particular priority, such as high, medium or low, in response to an associated memory queue reaching a selectable content level.
  • the interconnect networks provides a back pressure signal to the internal communication channels, wherein the back pressure signal contains the availability status for each of the internal communication channels.
  • a further enhancement of this feature utilizes communication bits, initially reserved for a destination address or handle, associated with a particular internal communication channel, to transfer the back pressure / availability status from a local interconnect network to an associated local communication interface.
  • the invention provides enhanced error correction.
  • the local interconnect network includes elements for generating a redundant version of information transferred from the local interconnect network to the local communication interface. If error detection elements detect an anomaly in transferred information, error correction elements can recover an error-free version of information from the redundant version.
  • the communication node includes control elements for deactivating those elements, be they line cards or interconnect elements, causing the detected anomalies.
  • the communication node provides improved fault-tolerance by deactivating failed line cards or interconnect elements, without compromising the speed with which information is transferred through the node.
  • the communication node generally, and the interconnects specifically, may also include circuit protection elements for enabling the hot replacement of failed components, while the communication node continues to transfer information.
  • the interconnect networks transfer information internally as information cells, wherein each cell includes groups of information words, and each group of information words is transferred by way of a different internal communication channel.
  • the local interconnect network generates the redundant version by performing a bit-by-bit "exclusive or” operation on pairs of groups of information words, prior to the pair being transferred to the local communication interface.
  • the local interconnect network also transfers the "exclusive or” version of the pair to the local communication interface.
  • the local communication interface can reconstruct an error-free version of the anomalous member by performing an "exclusive or” operation between the non-anomalous member and the "exclusive or” version of the transferred pair.
  • the interconnect networks can include elements for "clumping" or combining a plurality of information cells, and for transferring those clumped cells substantially simultaneously.
  • the interconnect networks also include elements for appending "dummy" cells to fill in a partial clump prior to the clump being transferred.
  • an interconnect network implements the clumping feature by employing storage queues associated with the transceivers.
  • the storage queues intermediately store groups of information words to be transferred.
  • the interconnect networks can further include detection elements for detecting when groups of words of a plurality of information cells to be included in a clump are stored in a queue, and transfer elements for substantially simultaneously transferring the clumped information cells by coupling the groups of words to transceivers.
  • the invention provides a substantially non-blocking feature.
  • the expanded interconnect network can include a forwarding array for storing data indicative of an unblocked local path through the expanded interconnect network.
  • the expanded interconnect network can use at least a portion of the destination address of a group of words of an information cell as a pointer into the forwarding array to select an unblocked path.
  • the expanded interconnect network employs a plurality of forwarding arrays, each storing data indicative of a segment of an unblocked path through the expanded interconnect network. Further, the expanded interconnect network can use successive portions of the destination address as pointers into each of the forwarding arrays to select each segment of an unblocked path.
  • the transfer elements of the local interconnect network and the transfer elements of the expanded interconnect network are essentially identical, and therefore, interchangeable.
  • the transfer elements can include a mode selection feature for selecting whether the element is to be used in a local mode or in an expanded mode. Such a feature provides substantial cost savings over prior art systems.
  • the invention includes methods corresponding to the above described apparatus.
  • FIGURE 1 depicts a plurality of communication networks interfacing through a communication node employing interconnection networks according to the invention
  • FIGURE 2 is block diagram of a communication node of the type depicted in FIGURE 1 and incorporating interconnect networks according to the invention
  • FIGURE 3 is a more detailed interconnection diagram of an exemplary local line card module and local interconnect module of the type depicted in FIGURE 2;
  • FIGURE 4 is a table illustrative of the interconnections between a typical line card module and an interconnect module of the types depicted in FIGURES 2 and 3;
  • FIGURE 5 is a more detailed block diagram showing interconnections between an exemplary local interconnect module and an expanded interconnect module of the types depicted in FIGURE 2;
  • FIGURE 6 is a table detailing the communication channel connections between an exemplary local interconnect module and the expanded interconnect module
  • FIGURE 7 is a table further specifying the connection of communication channels between potential local interconnect modules and the expanded interconnect module;
  • FIGURE 8 is a functional block diagram of a local interconnect logical plane of the type depicted in FIGURE 3 ;
  • FIGURES 9A and 9B are conceptual illustrations of high- and low-priority queues
  • FIGURES 10A and 10B are flow-charts depicting the operation of the queues of
  • FIGURES 9 A and 9B; and FIGURE 11 is functional block diagram of an expanded interconnect logical plane of the type depicted in FIGURE 5.
  • the invention is directed to an interconnection network in a communication node.
  • Communication nodes are junctions for transferring communication signals between a plurality of sources.
  • communication nodes may be required to interface with systems employing a variety of communication protocols and operating at differing information transfer speeds.
  • Prior art systems typically require maintenance of a plurality of networks, each being capable of interfacing with a particular type of source.
  • a communication node embodying features of an illustrative embodiment of the invention, can process information entering the node at a variety of speeds and formatted pursuant to a plurality of protocols.
  • information can enter and leave the communication node at OC48, OC12/STM4, OC3/STM1, DS3 and E3 speeds.
  • information can enter and leave the node in IP- or ATM-based formats.
  • a communication node employing interconnection networks employs a modular design.
  • the modular design enables a service provider to change the number of communication channels by adding or subtracting physical proximately located modules to or from the communication node.
  • the modules include a plurality of I/O interfaces coupled to an associated interconnection network.
  • the communication node employs a two-level interconnection network modularity; a local level and an expanded level.
  • a plurality of local interconnection network modules couple to an expanded interconnection network, also preferably located proximate to the local interconnection modules.
  • an expanded interconnection network also preferably located proximate to the local interconnection modules.
  • FIGURE 1 illustrates a typical deployment of a communication node 100 employing interconnection networks according to the invention.
  • the communication node 100 can support, among other systems, SONET rings 10, WWW server farms 20, dial up IP systems 30, IP over CATV/xDSL systems 40, IP over FR networks 50, multi-service ATM networks 60, ATM transports 70, and internet WDM systems 80.
  • Multi-service ATM access 60 can include CBR circuits 61, voice over ATM circuits 62, internet access 63, FR over ATM data 64, and LANs 65.
  • FIGURE 2 shows a block diagram of a communication node 100 incorporating interconnect networks according to an illustrative embodiment of the invention.
  • the communication node 100 employs a two-level interconnect architecture.
  • the node 100 includes one or more local line card modules 102-116, each having first- level associated local interconnect modules 118-132.
  • the communication node 100 further includes a second-level expanded interconnect module 134, sometimes referred to as a hyperconnect fabric.
  • the illustrated expanded interconnect module 134 includes three expanded interconnect boards 136-140. The boards are printed circuit boards holding integrated circuits and other components.
  • the local line card modules 102-116 are essentially identical to each other.
  • the local interconnect modules 118-132 are also essentially identical to each other. Accordingly, for the most part, the following description focuses on an exemplary local line card module 102 and an exemplary local interconnect module 118; that description being equally applicable to the remaining local line card modules 104-116, and the remaining local interconnect modules 120-132.
  • the local line card module 102 transfers information into and out of the communication node 100, by way of a plurality of I/O interfaces.
  • I/O interfaces can be, for example, IP or SONET/SDH ports that accept an OC48 data stream.
  • optional access modules 162-180 can be employed to provide OC12/STM4, OC3/STM1, DS3 and E3 ports.
  • Access module 162 multiplexes input data streams into an OC48/STM16 uplink to local line card module 102.
  • Line card module 102 couples information to an associated local interconnect module 118 by way of a plurality of Gigabit Ethernet connections 142.
  • the local interconnect module 118 transfers information between the I/O channels of local line card module 102.
  • a feature of the local line card module 102 is that it supports a number of datalink layer encapsulations, implemented by a flexible encapsulation/decapsulation mechanism.
  • the decapsulation mechanism is adaptable to accommodate emerging encapsulations.
  • local line card module 102 supports IP over ATM over SONET/SDH; IP over PPP over SONET/SDH; IP over FR over SONET/SDH; IP over PPP over FR over SONET/SDH; IP over PPP over ATM over SONET/SDH; IP over MPLS over SONET/SDH; IP over SNAP 802.2; and IP over Ethernet 2.0.
  • Line card module 102 also supports FRAME Relay DTE.
  • a further feature of the illustrated communication node 100 is that it is dynamically bandwidth scalable. More particularly, according to one embodiment, the communication node 100 can include up to seven additional local line card modules 104-116, coupled to associated local interconnect modules 120-132 by way of Gigabit Ethernet connections 144-160. The expanded interconnect module 134 transfers information between local interconnect modules 118-132 by way of a plurality of
  • FIGURE 3 is a more detailed schematic block diagram of an illustrative embodiment 200 of the communication node 100, but having only a single local line card module 102 coupled to a single local interconnect module 118. Since there is only a single local line card module 102, there is no need for the communication node 200 to include the second-level expanded interconnect module 134 of the type shown in FIGURE 1.
  • the local line card module 102 includes eight local line cards 202-216.
  • Local line cards 202-216 are printed circuit boards holding integrated circuits and other components. Each line card 202-216 has six internal (I/O) ports 202a-202f, and an external SONET I/O port 202g.
  • Line card 202 couples information between external I/O port 202g and internal I/O ports 202a-202f.
  • External I/O port 202g couples information into and out of the node 200, and the internal I/O ports 202a-202f connect with up to forty-eight internal communication lines 217 and couple information between the local line card module 102 and the local interconnect module 118.
  • each internal I/O port a-f includes a Gigabit Ethernet transceiver, providing a Gigabit Ethernet input channel and a Gigabit Ethernet output channel.
  • the input and output channels provide 10-bits of information.
  • transceiver as used throughout this description, is also intended to encompass structures including separate receivers and transmitters.
  • the external I/O port 202g is preferably software configurable for either SONET or SDH operation.
  • physical interfaces are software configurable for OC48 or STM16.
  • SONET and SDH PAMS may be freely intermixed within access module 142.
  • a fully loaded local line card module 102 can have up to eight external SONET/SDH I/O ports and forty-eight corresponding internal I/O ports.
  • the local interconnect module 118 includes three identical interconnect boards 218-222.
  • the interconnect boards 218-222 are printed circuit boards holding integrated circuits and other components. Each board 218-222 is logically subdivided into two essentially identical planes.
  • interconnect board 218 includes logical planes 218a and 218b;
  • interconnect board 220 includes logical planes 220a and 220b; and
  • interconnect board 222 includes logical planes 222a and 222b.
  • the communication node 200 transfers information through the interconnect boards 218-222 by way of Application Specific Integrated Circuits (ASICs) 224-228. Each ASIC 224-228 logically includes an a-half and a b-half.
  • ASICs Application Specific Integrated Circuits
  • the logical a-half services the logical a-plane of a particular interconnect board 218-222, while the logical b-half services the logical b-plane of the particular interconnect board 218-222.
  • ASIC 224a services logical plane 218a
  • ASIC 224b services logical plane 218b.
  • FIGURES 2 and 3 is described with respect to particular physical groupings of line cards 202-216 and local interconnect boards 218- 222, and particular logical divisions of local interconnect planes 218a, 218b, 220a, 220b, 222a and 222c.
  • the number of line card modules 102-116 and associated local interconnect modules 118-132 can be varied. Moreover, the number of local line cards included in a local line card module 102-116 can also be varied. Further, the number of local interconnect boards included in a local interconnect module 118 and the number of extended interconnect boards included in an extended interconnect module 134 can be varied. All logical divisions can also be varied.
  • each local interconnect board 218-222 includes sixteen internal communication ports (eight associated with each logical plane), and sixteen expanded communication ports (eight associated with each logical plane).
  • board 218 will be discussed in detail.
  • Boards 220 and 222 have a similar construction and operation.
  • local interconnect board 218 has eight internal communication ports 0a-7a, associated with ASIC 224a, and eight internal communication ports 0b-7b, associated with ASIC 224b.
  • Local interconnect board 218 also includes eight expanded communication ports 8a- 15a, associated with ASIC 224a, and eight expanded communication ports 8b- 15b, associated with ASIC 224b.
  • Each internal and expanded communication port includes an Ethernet transceiver providing a Gigabit Ethernet input channel and a Gigabit Ethernet output channel.
  • Each internal communication port 0a-7a and 0b-7b couples to an internal communication port a-f of a line card 202-216, and transfers information between the local line card module 102 and the local interconnect module 118.
  • each internal communication port a-f of line cards 202-216 couples to an internal communication port 0a-7a and 0b-7b of one of the interconnect boards 218-222.
  • FIGURE 4 shows a table 300 illustrative of the interconnections between the internal communication ports of line cards 202-216 and the internal communication ports of interconnect boards 218-222.
  • Column 302 shows the line card designation numbers 202-216.
  • Column 304 lists sets of ordered pairs in the form (interconnect plane #, port #), wherein those ordered pairs identify which internal port 0a-7a and 0b-7b of interconnect module 118 couples to each internal port a-f of line cards 202-216.
  • the first line of table 300 indicates that line card 202, internal port a, couples to internal port 0a of interconnect plane 218a.
  • internal ports b, c, d, e, and f of line card 202 couple to internal ports 0b of plane 218b, 0a of plane 220a, 0a of plane 220b, 0a of plane 222a, and 0b of plane 222b, respectively.
  • the internal communication ports of line card 202 couple to the "zero numbered" ports of the interconnect planes 224a, 224b, 226a, 226b, 228a and 228b.
  • the internal communication ports of line card 204 couple to the "one numbered" ports (e.g. la).
  • Those of line card 206 couple to the "two numbered" ports (e.g. 2a).
  • Those of line card 208 couple to the "three numbered" ports (e.g. 3a), and so on.
  • the communication node 200 transfers information from the local interconnect module 118 to associated local line card module 102 in 64-byte cells.
  • the 64-byte cell is subdivided into 16-byte groups.
  • Logical plane 218a transfers 8-words of 2-bytes each.
  • Logical planes 218b, 220a and 220b do the same.
  • the two least significant bytes (LSBs) of the first and third 16-byte groups i.e., the groups transferred by logical planes 218a and 220a
  • the two LSBs of the second and fourth 16- byte groups i.e., the groups transferred by logical planes 218b and 220b
  • Board 222 provides error correction and redundancy information.
  • logical plane 222a provides a bit-by-bit "exclusive or” ( ⁇ ) between the information transferred on logical plane 218a and logical plane 220a.
  • Logical plane 222b provides a bit-by -bit "exclusive or” between the information transferred on logical plane 218b and logical plane 220b.
  • a byte contains 8-bits and a word contains 2-bytes.
  • alternative byte and word conventions may be employed.
  • Each logical plane 218a, 218b, 220a and 220b receives a 16-byte group.
  • Logical plane 222a receives the "exclusive or” of planes 218a and 220a
  • logical plane 222b receives the "exclusive or” of logical planes 218b and 220b.
  • Bytes 2-16 of logical planes 218a, 218b, 220a and 220b provide the transferred data.
  • X represents an XOFF from the line card to the local interconnect 118.
  • Information cells pass through the local interconnect 118 on separate planes 218a-222b, but with some discrepancies.
  • the a-planes contain the line card destination addresses.
  • the b-planes allow for 14-bits of extra "payload” data carried through untouched.
  • the payload byte above typically has its most significant bit (bit "P") set as parity for words 0 and 1 together. Even on plane 222a, the "P" bit covers the 15-bits which precede it, rather than the parity across planes 218a and 220a.
  • the line card destination address is an address or handle (global to the node 100 system wide) which specifies the destination line card (for unicast information) or line card set (for multicast information) to which the information cell is to be transferred.
  • TABLE 3 depicts a preferred destination address format for locally transferred unicast information
  • TABLE 4 depicts a preferred destination address format for locally transferred multicast information.
  • Byte-0 of TABLE 3 provides an "APS" bit, a Line Card Module designation field and a Line Card designation field.
  • FIGURE 3 there are eight potential destination line cards 202-216. Each one is assigned a 3-bit code 000-111. Bits 3, 4 and 5 of byte-0 of TABLE 3 provide this code.
  • a communication node 100 according to the illustrated embodiment, has eight potential line card modules 102-116. Each line card module 102- 116 is assigned a 3-bit code 000-111. Bits 0, 1 and 2 of byte-0 of TABLE 3 provide this code. Bytes 2-15 provide the transferred data. The "P" bit provides parity across byte 1.
  • unicast information cells travelling to a single destination line card have the "Pri" bit of TABLE 3 set to one for high-priority traffic. If the "APS" bit of TABLE 1 is set, the cell is sent to both the designated line card (n) and the (n+1) line card.
  • the multicastID is an address into a 16 k x 9-bit RAM 748 of FIGURE 8.
  • This 14-bit address retrieves a 9-bit value.
  • the least significant 8-bits are a mask with each bit representing a line card 202-216.
  • the most significant bit represents priority; set to a logical one for high priority or a logical zero for low priority. If the node includes more than one local line card chassis 102, then each bit of the 8-bit mask represents a particular local line card chassis.
  • the communication node 200 transfers each 16-byte group over a different internal communication channel.
  • each line card 202-216 has an associated address, and information enters line card 202 by way of external port 202g. Assume further that the entering information has a destination address of line card 208.
  • internal port 202a couples a first 16-byte group to internal port 0a of interconnect board 218.
  • Internal port 202b couples a second 16-byte group to internal port 0b of board 218.
  • Internal port 202c couples a third 16-byte group to internal port Oa of board 220, and internal 202d couples a fourth 16-byte group to internal port Ob of board 220.
  • Internal port 0a of board 218 couples the first 16-byte group to ASIC 224a.
  • Internal port Ob of board 218 couples the second 16-byte group to ASIC 224b.
  • Internal port 0a of board 220 couples the third 16-byte group to ASIC 226a, and internal port Ob of board 220 couples the fourth 16-byte group to ASIC 226b.
  • ASIC 224a processes the line card destination address and directs the first 16- byte group to internal port 3a of board 218.
  • ASIC 224b processes the line card destination address and directs the second 16-byte group to internal port 3b of board 218.
  • ASIC 226a processes the line card destination address and directs third 16-byte group to internal port 3a of board 220.
  • ASIC 226b processes the line card destination address and directs the fourth 16-byte group to internal port 3b of board 220.
  • Board 222 generates a bit-by-bit "exclusive or" between the first and third groups, and between the second and fourth groups.
  • ASIC 228a processes the line card destination address and directs the "exclusive or” combination of the first and third groups to internal communication port 3a of board 222
  • ASIC 228b processes the line card destination address and directs the "exclusive or” combination of the second and fourth groups to the internal communication port 3b of board 222.
  • Board 222 in turn couples the first, second, third and fourth groups to destination line card 208, internal ports 208a-208d, respectively.
  • board 222 couples the "exclusive or” version of the groups of bytes to internal ports 208e and 208f.
  • Destination line card 208 then performs various types of error checking, such as plain parity, 8B10B disparity and CRC across multiple cells.
  • the line card module 102 can determine whether a detected error is due to a failed line card 202-216 or a failed interconnect board 218-222. In the case where the error is due to a failed line card, the interconnect module removes that card from operation. In the case where the detected error is due to a failed interconnect card 218-222, the line card detecting the error can signal the error prone interconnect board 218-222 to take itself off line. So as not to compromise bandwidth, interconnect board 222 can automatically take the place of either interconnect board 218 or 220, until the failed board is replaced. According to a further feature, the failed board can be hot-swapped.
  • the communication node 100 is dynamically bandwidth scalable to include additional line card modules 104-116, having additional associated local interconnect chassis 120-160.
  • the modular construction of the line card modules 102-116, along with the modular construction of the local interconnect modules 118-132, in combination with the expanded interconnect module 134 provides the scalable feature. More specifically, as indicated in FIGURE 2, according to the illustrated embodiment 100, a first mechanical chassis 101 houses up to eight local line card modules 102-116. Similarly, a second mechanical chassis 103 houses up to eight local interconnect modules 118-132. Further, a third mechanical chassis 105 houses the extended interconnect network 134.
  • the first mechanical chassis 101 electrically couples to the second mechanical chassis 103 by way of communication channels 142- 160.
  • the second mechanical chassis 103 electrically couples to the third mechanical module 105 by way of communication channels 170.
  • the communication node 100 employs connectors designed for "hot-swapping" at module interfaces. Those connectors enable the local line card modules 102-116, the local interconnect modules 118-132, and the expanded interconnect boards 136-140 to be connected and unconnected (i.e., "hot-swapped") from their respective mechanical chassis 101, 103 and 105, while the communication node 100 is powered and operating transferring information.
  • FIGURE 5 is a more detailed schematic block diagram 400 showing interconnections between an exemplary local interconnect module 118 and an expanded interconnect module 134.
  • the embodiment of FIGURE 4 is illustrative of the dynamic bandwidth scalability of the communication node 100.
  • the local interconnect module 118 is essentially identical to optional local interconnect modules 120-132. Accordingly, the following illustrative description focuses on the interconnections between local interconnect module 118 and expanded interconnect module 134. However, each additional local interconnect module 120-132 interfaces to expanded interconnect module 134 in a like manner.
  • the local interconnect module 118 includes three interconnect boards 218-222.
  • Each board 218-222 logically subdivides into an a-plane and a b-plane.
  • the a-plane and the b-plane are essentially identical and share an ASIC, which is also logically subdivided into an a-half and a b-half.
  • Each ASIC 224-222 provides, among other functions, routing and switching between the internal communication channels associated with the particular interconnect board 218-222.
  • each ASIC 224-226 may also provides storage queues, a control processor, control registers and status registers.
  • Each local interconnect board 218-222 includes sixteen internal I/O ports 0a-7a and 0b-7b.
  • the internal I/O ports 0a-7a and 0b-7b provide Gigabit Ethernet interfaces. As shown in FIGURE 3, those Gigabit interfaces couple to the internal communication channels of an associated local line card module 102.
  • Each local interconnect board 218-222 also provides sixteen expanded I/O ports 8a- 15a and 8b- 15b.
  • Each expanded I/O port 8a- 15a and 8b- 15b provides a Gigabit input channel and a Gigabit output channel.
  • the expanded I/O ports 8a-15a and 8b-15b couple information between the local interconnect module 118 and the expanded interconnect module 134.
  • the expanded interconnect module 134 includes three essentially identical expanded interconnect boards 136-140.
  • Each board 136-140 includes, among other components, one hundred and twenty-eight Gigabit Ethernet transceivers.
  • Each board 136-140 also includes four ASICs 402-408, 410-416 and 418-424, respectively.
  • ASICs 402-424 are essentially identical to ASICs 224-228. However, ASICs 404-424 are mode selected to operate in an expanded interconnect mode, rather than the local interconnect mode of ASICs 224-228. As in the case of ASICs 224-228, ASICS 404-424 each logically subdivides into an a-half and a b-half.
  • Each half includes sixteen Gigabit Ethernet I/O ports, wherein each port includes a Gigabit input channel and a Gigabit output channel.
  • Each of the sixteen Gigabit Ethernet ports couple to a Gigabit transceiver on the extended interconnect board.
  • board 136 of FIGURE 5 includes ASICs 402-408.
  • ASIC 402 is subdivided into to logical halves 402a and 402b.
  • ASIC 404 is subdivided into logical halves 404a and 404b;
  • ASIC 406 is subdivided into logical halves 406a and 406b;
  • ASIC 408 is subdivided into logical halves 408a and 408b.
  • ASIC 402 includes Gigabit Ethernet ports 0a-15a, on half 402a, and 0b- 15b on half 402b. Ports 0a-15a couple to transceivers 0-15 on board 136, and ports 0b-15b couple to transceivers 16-31.
  • Gigabit ports 0a-15a and 0b- 15b of ASICS 404-408 successively couple to remaining transceivers 32-127.
  • Gigabit I/O ports of ASICs 410-416 and 418- 424 couple to one hundred and twenty-eight transceivers of boards 138 and 140, respectively, in an identical fashion to that described with respect to ASICS 402-404 on board 136.
  • FIGURE 6 is a table 500 illustrating the communication paths between local interconnect module 118 and expanded interconnect module 134.
  • Column 502 specifies the local interconnect port designation.
  • Column 504 specifies the communication paths between local interconnect board 218 and expanded interconnect 134.
  • column 506 specifies the communication paths between local interconnect board 220 and expanded interconnect 134;
  • column 508 specifies the communication paths between local interconnect board 222 and expanded interconnect 134.
  • the communication channels of board 218 couple to board 136.
  • Those of board 220 couple to board 138, and those of board 222 couple to board 140.
  • FIGURE 7 is a table 600 depicting the expanded interconnect port assignments for each of the potential local interconnect modules 118-132 of FIGURE 1.
  • Column 602 identifies the particular local interconnect modules 118-132.
  • Column 604 identifies the port assignments for the extended interconnect module 134.
  • local interconnect module 120 couples to the one and nine ports; module 122 couples to the two and ten ports; module 124 couples to the three and eleven ports; module 126 couples to the four and twelve ports; module 128 couples to the five and thirteen ports; module 130 couples to the six and fourteen ports; and module 132 couples to the seven and fifteen ports.
  • TABLE 5 specifies a preferred format for the destination address for unicast information transferred from a local interconnect modules 118-132 to the expanded interconnect module 134.
  • TABLE 6 specifies a preferred format for the destination address for multicast information transferred from a local interconnect modules 118-132 to the expanded interconnect module 134.
  • bit 7 of the most significant byte is the parity bit, which represents parity across the previous 15-bits of the destination address.
  • Bit 5 is the "Valid” bit.
  • the “Valid” bit is set if the destination address is valid.
  • Bit 4 is the “Clump” bit.
  • the “Clump” bit is set if there is a valid combination or clump of cells. Clumping is a feature of the invention employed for eliminating the need for reordering transferred information subsequent to transfer. As discussed in further detail below, with reference to FIGURE 7.
  • Bit 7 is the “Pri” bit.
  • the “Pri” bit is set to one for high- priority traffic.
  • Bit 7 of the LSB is the "APS" bit.
  • the "APS" bit is simply passed through the expanded interconnect 134 to local interconnects 118-132.
  • Bits 3-5 of the LSB provide the 3-bit designation 000-111 for the destination line card module.
  • Bits 0-3 of the LSB provide the 3-bit designation 000-111 for the particular line card.
  • P 1 Mu ticastID (not 0 x 3FFF)
  • the "Multicast ID" of TABLE 6 is passed on to the expanded interconnect module 134 to be translated. If the payload data portion of the cell is 0x3 FFF, the information cell is considered to be invalid.
  • TABLES 7 and 8 depict a preferred format for unicast and multicast destination addresses, respectively, for information cells transferred from the expanded interconnect module 134 to a local interconnect modules 118-132.
  • the "P" bit provides parity across the destination address.
  • the "Valid” bit is set if the destination address is valid.
  • the "Pri” bit is set for high-priority traffic.
  • the "APS” bit is set if the cell is to be sent to both the designated line card (n), and the (n+1) line card. Bits 0-3 provide the designation code for the line card to which the cell is sent.
  • FIGURE 8 is a functional block diagram of an illustrative local interconnect board of FIGURE 5.
  • the board 218 includes an ASIC 224 of the type employed in the local interconnect module 118.
  • ASIC 224 is identical to ASICs 402-424 employed in the expanded interconnect module 134.
  • all of the local interconnect boards are preferably identical, for the purpose of the following discussion it is assumed that board 218 is an exemplary local interconnect board of local interconnect module 118, and interfaces with local line card module 102 of FIGURE 2.
  • the ASIC 224 need not be identical to ASICs 404-424.
  • the interconnect board 218 includes Gigabit transceiver sets 704 and 708, memory elements 710, controller 712 and status and control registers 714.
  • Gigabit transceiver set 708 provides Gigabit I/O ports 0a-7a and 0b-7b, which couple to the internal communication channels of a local line card module 102 of FIGURE 3.
  • Gigabit transceiver set 704 provides Gigabit I/O ports 8a- 15a and 8b- 15b, which couple to the expanded communication channels of the expanded interconnect module 134, shown in FIGURE 5.
  • Each transceiver of sets 704 and 708 couples to the ASIC 224 by way of associated input and output shift and hold registers. More specifically, transceivers of set 704 couple to input shift and hold registers 714 by way of lines 716 and output shift and hold registers 718 by way of lines 720. Transceivers of set 704 couple to input shift and hold registers 722 by way of lines 724, and output shift and hold registers 726 by way of lines 728.
  • the ASIC 224 also includes a dual-port RAM 730 for storing various stacks and queues 731 associated with flow control information.
  • Flow status 733 stores an availability status, regarding the availability of a particular line card to receive information.
  • RAM 730 intermediately stores information being transferred through the board 218.
  • Shift and hold registers 714 and 716 couple to the dual-port RAM 730 by way of lines 732 and 734, respectively.
  • Shift and hold registers 722 and 726 couple to the dual-port RAM 730 by way of lines 736 and 738.
  • the dual-port RAM 730 also couples to destination stack 740 by way of lines 742.
  • the ninety-six destination queues 740 intermediately store addresses representative of where particular data is stored in RAM 730.
  • the queues 740 preferably employ a plurality of stacks for ease of addressing. However, other storage structures can be employed. As discussed above in the Summary of the Invention, and as discussed in further detail below, according to a preferred embodiment, the invention employs a plurality of memory storage queues / buffers to aid in the efficient transfer of information. It should be noted that the terms queue and buffer are used interchangeably.
  • the dual-port RAM 730 provides an output queue for each transceiver of sets 704 and 706. More specifically, information cells coupled into board 218 to be transferred to a line card 202-204 of local interconnect 102, are first written into buffer memory at an address which is written into an output queue. Free list memory 742 provides a list of available buffer memory addresses.
  • Reference counter 744 for each of the 1536 buffers in the dual port RAM 730.
  • Reference counter 744 contains the number of output queues to which the contents of the respective buffers are to be sent.
  • a reference counter 744 decrements in response to information being read from an associated buffer. When the reference counter reaches zero, the address of the buffer is returned to free list 743. In this way, the ASIC 224 can track the available buffer locations associated with each transceiver.
  • Information written to buffer memory is subsequently transferred to one of the output shift and hold registers 720 or 728, and held there until an internal time slot arrives in which the destination address lookup can be performed, the read from the free list memory 742 can be performed, the write to the buffer memory can be performed, and the write to the output queue can be performed.
  • queues 731 can include QoS queues.
  • the QoS queues such as those conceptually illustrated in FIGURES 9A and 9B, can have multiple watermark levels; those levels corresponding to differing priorities.
  • high- priority queue 900 of FIGURE 9A can have two watermarks 906 and 908.
  • queue 900 reports its status as "stop-none," indicating the I/O channel associated with queue 900 is ready to receive information of any priority.
  • queue 900 reports its status as "stop- low” 1002, indicating the I/O channel associated with queue 900 is ready to receive information having a "medium” priority or higher, as indicated at 1004.
  • the queue 900 is filled up to level 906, it reports its status as "stop-all,” as shown at 1006. As shown at 1008, this indicates that its associated I/O channel is unavailable. If the "Stop-Low" watermark 908 of queue 900 has not been reached, as indicated at 1006, it is available to receive information of any priority.
  • Low-priority queues can include three watermarks 918, 920 and 922.
  • queue 910 reports a status of "Stop-None" in range 912.
  • queue 914 and 101B it reports a status of "Stop-Low” in range 914.
  • queue 918 and 1020 it reports a status of "Stop-Medium” in range 916, and as shown at 1022 and 1024, queue 910 reports a status of "Stop- All" subsequent to reaching watermark 918.
  • High-priority queues such as queue 900, enable associated line cards to pass low- and medium-priority traffic, while not allowing low-priority traffic of one line card to strangle medium-priority traffic of a different line card.
  • Low-priority queues such as queue 910, enable associated line cards to pass low-, medium- and high-priority traffic, while not allowing low-priority and medium- priority traffic of one line card to strangle high-priority traffic of a different line card. It also prevents low-priority traffic of one line card from strangling medium- and high- priority traffic of a different line card.
  • the dual-port RAM 730 preferably provides storage for sixty-four low-priority unicast queues; one for each possible local line card in the communication node 100.
  • the RAM 730 also provides storage for sixteen high-priority unicast queues; one for each line card of its local interconnect module, one for each potential additional local interconnect module, and one extra queue.
  • Multicast traffic preferably employs four low-priority and four high- priority queues.
  • each plane of the expanded interconnect 134 employs eight high- priority unicast queues; one for each potential local interconnect module 118-132.
  • Each expanded interconnect logical plane also employs eight high-priority and eight low- priority multicast queues; again, one for each potential local interconnect destination module 118-132.
  • the status provides information regarding the availability of each line card 202-216 to receive information of varying priority levels.
  • Another feature of the illustrated embodiment of the invention is the way in which the node 100 passes the flow control status (sometimes referred to as back pressure status) between the expanded interconnect module 134 and each of the line cards of the local interconnect modules 118-132.
  • the invention utilizes bits of the information cell, previously reserved for the destination address. These bits are indicated in TABLE 1 as the "Flow Control" words on the b-channels.
  • Flow control information is passed between the local interconnect modules 118- 132 and the expanded interconnect module 134 using the least significant word of the b- channel. These bits are included in the parity calculation of the parity bit in the primary channel's destination address word. This format is generally illustrated above in TABLE 1 , with respect to local interconnect plane 218a word 0, and local interconnect plane 218b word 0. This flow information is preferably not repeated on all links. As illustrated in TABLE 9 below, also with reference to local interconnect planes 218a and 218b and expanded interconnect board 136, flow control information is sent in a two- cell sequence. More particularly, column 1 of TABLE 9 lists the expanded interconnect port in
  • Type 0 and Type 1 identifies the information contained in the byte (e.g. if local interconnect port 8b receives a Type 0 byte, that byte contains Low6, Low5, Low4, Low3, Low2, Lowl and Low 0 flow control information).
  • Each of the Low0-Low63 bits are set if the corresponding low- priority queue is not full and thus, can receive data.
  • the High0-High7, MCHigh, and MCLow bids are set if the corresponding high-priority, multicast high- priority and multicast low-priority queues have space available for receiving information.
  • the watermark levels are programmed at a level that takes into account potential latency.
  • flow control latency takes four cells to stop incoming information
  • (4* 16)-4 locations should be reserved above the watermark to avoid overflow. This results from each of sixteen local input ports potentially aiming at the queue for four cell times. Thus, it would be draining out four information cells in that interval.
  • the space below the watermark level need only be (1 *flow control latency) to avoid underflow.
  • the interconnect networks support Constant Bit Rate (CBR), Variable Bit Rate-Real-Time (VBR-rt), Variable Bit Rate-Non-Real-Time (VBR-nrt), and Unspecified Bit Rate (UBR) QoS categories.
  • CBR Constant Bit Rate
  • VBR-rt Variable Bit Rate-Real-Time
  • VBR-nrt Variable Bit Rate-Non-Real-Time
  • URR Unspecified Bit Rate
  • the interconnect networks can operate as a class-based ATM switch. Thus, traffic is queued for transfer based on the service category of the virtual circuit. However, shaping and policing are performed on a per-virtual-circuit basis.
  • the interconnect networks also support QoS features for IP networks, such as the Differentiated Services Model.
  • a preferred embodiment of the invention employs "clumping" to increase the rate with which information can be transferred through the interconnect networks.
  • portions of communications can pass through an interconnect network at varying speeds, thus arriving at a common destination in a misordered fashion. Reordering information subsequent to transfer can waste valuable time, and has the potential for receiving out of order cells. Therefore, according to a preferred embodiment of the invention, the expanded interconnect network 134 includes elements for "clumping" or combining a plurality of information cells and for transferring the clumped cells substantially simultaneously.
  • the queue depth logic 746 detects when a group of four unicast information cells are available in a single queue. In response to detecting four unicast cells in a single queue, the queue depth logic 746 signals the dual-port RAM 731 working in conjunction with the destination stack 740 to transfer the detected four cell clump to shift and hold registers 726 for substantially simultaneous transfer via Ethernet transceivers 708. With the clump of cells being transferred together, they arrive at a destination within a close enough time proximity to avoid reordering. According to a further embodiment, a programmable wait timer begins decrementing upon the arrival of a first information cell to be included in the clump.
  • the timer If the timer expires prior to the complete formation of a clump, it triggers the cell(s) ready to be sent to be combined with 4-N invalid cells, where N is the number of cells which the clump is lacking.
  • Multicast cells are clumped together across paths. When multicast traffic is available to be sent on at least four different paths, it is considered available for transmission. As in the case of unicast traffic, a programmable wait timer on any given multicast queue can artificially render multicast traffic eligible. A programmable watermark threshold on multicast queues can also artificially render multicast traffic eligible. Whenever multicast traffic is eligible to be sent, "Qlnfo" cells are sent on the remaining links to or from the expanded interconnect 134.
  • the ASIC 224 also includes a translation memory 748. The translation memory
  • translation memory 748 provides storage for path segments through the expanded interconnect module 118, if such a module is included in the system.
  • translation memory 748 preferably contains nine logical storage areas; one for each local interconnect modules 102-116, and one for the expanded interconnect module 118.
  • the expanded interconnect storage area is configured as a bitmap of destination line cards and priority.
  • Destination address circuitry 750 accesses the translation memory 748, and the multicast bitmap register 752 receives the accessed information.
  • a feature of the invention is synchronization of the local interconnect boards of a particular local interconnect module with each other, and the synchronization of the expanded interconnect boards 138-140 with each other. Since each board is independent, although they derive their clock frequencies from the same source, some signals are employed to establish and maintain synchronization between boards. Slot synchronization forces essentially identical "time zero" references between boards, and thus planes, in a module. Cell synchronization enables a local interconnect module to set its slot zero reference such that its transmitted information cells can arrive at the expanded interconnect 134 at a safe time.
  • CPU synchronization enables certain CPU write operations to take place at the same time on all planes in a module.
  • Plane synchronization logic 752 provide signals 754 necessary to synchronize certain read and write operations between each plane of local interconnect module 118.
  • Control registers 758 provide flow control information by way of lines 760 to the other planes of local interconnect 118, if operating in local mode, or the additional expanded interconnect planes of module 134 if operating in expanded mode.
  • Plane to plane cell synchronization is attained by cellok inter-plane connections 761.
  • An asserted cellok signal 761 indicates that the corresponding plane has received a valid and error free cell header containing the 2-byte destination address.
  • each plane outputs 16 cellok signals 761 and inputs 32 cellok signals 761.
  • Each cellok output, N represents that both the a-ports and the b- ports have valid cell headers.
  • ASIC 224 includes mode select 756 for selecting whether ASIC 224 is to operate as a local interconnection circuit or as an expanded interconnection circuit.
  • ports 0a-7a and 0b-7b connect to local line cards and ports 8a-15a and 8b- 15b connect to expanded interconnect 134.
  • ports 0a-15a and Ob- 15b connect to local interconnect planes, such as 218-222 shown in FIGURE 5.
  • an important feature of the illustrated ASIC 224 is a "slot counter" contained in timers, counters, control registers 758.
  • the slot counter repeatedly counts from 0-15. Each port 01 -15a and Ob- 15b is assigned a slot count. Each time the slot count 0-15 matches a port number, a check is performed to determine if there is a cell to be transmitted out that port. If there is, the cell is copied from RAM 730 to shift and hold register 718 or 726 for transmission. If there is no cell to be transmitted, then a flow control cell is transmitted.
  • a common slot counter is employed for the a-ports and the b-ports.
  • board 218 also includes controller 712 and memory 710.
  • Memory 710 stores the control code for board 218. As such, it provides start up initialization of statuses, pointers and communication interfaces. Controller 712 provides a variety of conventional processor functions.
  • FIGURE 11 is described in terms of the a-plane 136a of expanded interconnect board 136 of FIGURE 5.
  • Plane 136a includes four ASICs 402a, 404a, 406a and 408a.
  • ASICs 402a-408a are essentially identical to ASIC 224 of FIGURE 8.
  • Each ASIC 402a-408a controls sixteen Gigabit Ethernet ports 1102-1108.
  • Ports 1102-1108 couple to ASICs 402a-408a by way of shift and hold registers, such as registers 714 and 718 of FIGURE 8.
  • Bus 110 couples interplane control signals, such as those generated by counters, timers and registers 758 and plane synchronization logic 753, to each ASIC 402a-408a.
  • Processor 1114 controls ASICS 402a-408a by way of bus 1112.
  • Processor 1114 includes a CPU module, DRAM, FPGA control and Ethernet control, much in the same way that memory 710, controller 712 and control and status registers 753 provide these functions for local interconnect board 218.
  • connections and circuit divisions referred to in the above description may be representative of both actual and logical connections or divisions.
  • FIGURE 11 shows a functional block diagram of a typical expanded interconnect logical plane illustrative of any of interconnect planes of FIGURE 4. It will thus be seen that the invention efficiently attains the objects set forth above, including providing dynamically bandwidth scalable interconnect network. Since certain changes may be made in the above constructions and the described methods without departing from the scope of the invention, it is intended that all matter contained in the above description or shown in the accompanying drawings be interpreted as illustrative and not in a limiting sense.

Abstract

An interconnect network for operation within communication node, wherein the interconnect network may have features including the ability to transfer a variety of communication protocols, scalable bandwidth and reduced down-time. According to one embodiment of the invention, the communication node includes a plurality of I/O channels for coupling information into and out of the node, and the interconnect network includes at least one local interconnect module having local transfer elements for transferring information between the plurality of I/O channels; and scaling elements for expanding the interconnect network to include additional local interconnect modules, such that information can be transferred between the local interconnect modules included in the interconnect network.

Description

AN INTERCONNECT NETWORK FOR OPERATION WITHIN A COMMUNICATION NODE
Technical Field The invention relates generally to communication nodes, and more particularly to an interconnect network for operation within a communication node.
Background of the Invention
Communication nodes, which act as junction points for communication signals transferred between a plurality of sources, are required to handle a variety of popular communication protocols, such as Integrated Services Digital Network (ISDN) protocol, Asynchronous Transfer Mode (ATM), and Internet Protocol (IP). ISDN, an early attempt at a multi-service architecture (i.e., an architecture capable of handling a variety of communication encapsulations), which is based on the telephone hierarchy, apportions bandwidth in 64 kilobits per second (Kbps) circuits. With local area networks (LANs) operating at 10 megabits per second (Mbps), ISDN has proved too slow. ATM is a packet switching protocol that was conceived as a transport mechanism for broadband ISDN. ATM transfers information in fixed-length packets called cells. The cells travel over virtual connections (VCs) between communication nodes that are established prior to each communication session. The combination of fixed cell formats and VCs renders ATM a faster alternative to ISDN. Additionally, ATM handles bursts of data traffic more efficiently than time division multiplexing (TDM) and provides high-quality voice and video support.
The popularity of the World Wide Web (WWW) has encouraged the use of IP. As a result, low-cost, distance-insensitive IP-based transport has become an attractive alternative to leased lines and frame relay (FR). Additionally, Internet Service Providers (ISPs) have become serious contenders for enterprise traffic.
Unfortunately, all of the information transfer protocols have drawbacks. Specifically, ISDN provides a relatively low-speed access solution. ATM supports frame relay, virtual private networks (VPNs), circuit emulation, private branch exchange (PBX) interconnects and quality of service (QoS), but does not mesh easily with existing data protocols. IP supports applications such as Internet Access and VPNs, for which cost connectivity is important. However, IP has yet to demonstrate industrial-strength reliability. As a result, full service providers find it necessary to maintain parallel switching networks. Because such parallel networks require maintenance and service of a variety of devices such as, voice switches, frame relay switches, ATM switches, routers, add/drop multiplexers, and digital cross-connects, they have a high associated capital equipment cost.
Conventional communication nodes also have a variety of drawbacks. For example, conventional communication nodes fail to provide sufficient ease of scalability. Typically, conventional switching nodes provide a switching/routing network having a fixed bandwidth. However, as enterprises grow, their needs also grow. But, the fixed bandwidth switching routing network of conventional technology requires enterprises to predict such growth and purchase systems having sufficiently large bandwidth up front; thereby compounding the challenge of maintaining parallel networks. Another drawback of conventional systems is reliability. Because conventional systems fail to provide a single switch/routing network that can operate on a variety of protocols, today's Giga Points-of-Presence (GigaPoPs) and Access PoPs are a complex and expensive aggregation of core routers connecting smaller Access PoPs to the core transport capacity. These structures are fragile, with frequent service outages due to performance limitations and equipment failures. Enterprises cannot afford to be exposed to significant down time due to failures or updates associated with conventional technology.
Because the switching/routing networks of conventional systems are typically designed to operate under the constraints of a particular protocol, they lack the flexibility to adapt to emerging technologies, employing new communication protocols. As discussed above, different protocols provide different QoS features. Thus, another drawback of a network operating under the constraints of a single protocol is that a service provider cannot offer varying grades of service to users having differing priority requirements; thus causing service providers to forego a potentially significant source of revenue. Accordingly, it is an object of the invention to provide an interconnect network that enables a multi-service communication node to handle a variety of communication protocols, without requiring the maintenance of costly parallel networks. It is a further object of the invention to provide an interconnect network that enables a communication node to adapt to communication protocols employed by emerging technologies.
Another object of the invention is to provide a scalable interconnect network enabling bandwidth scaling of a communication node to fit the needs of providers having varying bandwidth requirements.
An additional object of the invention is to provide a fault-tolerant interconnect network capable of repair and update, without causing down-time or compromising operation of the communication node. These and other objects of the invention will be apparent with respect to the following description of the invention.
Summary of the Invention
The invention is directed to communication nodes. More particularly, it is directed to interconnect networks in communication nodes. According to one embodiment of the invention, a communication node includes interconnect networks that enable the node to transfer a variety of communication protocols. According to a further embodiment, an interconnect network according to the invention enables a communication node to handle ATM and IP Packet-over-SONET protocols with the same hardware. An interconnect network according to an additional embodiment of the invention also enables a communication node to provide Frame Relay Data Terminal Equipment (DTE) and Multiprotocol Label Switching(MPLS) functionality. An interconnect network according to another embodiment of the invention enables a communication node to act as both a native ATM switch and a native IP router, operating at line speeds up to at least as high as 2.488 Gps (OC48c/STM16c).
According to additional features, the invention can provide improved reliability. By way of example, according to one embodiment, the invention provides Automatic Protection Switching (APS), wherein Open Systems Interconnection (OSI) Layer 2 and Layer 3 information is mirrored to provide rapid APS switchover. Additionally, system modules can be hot-swappable, and designed so that single component failures do not lead to total node failure. According to another embodiment, the communication node is packaged in a scalable set of modules. OC48 line cards and Gigabit Ethernet modules populate a local communication interface module including a local interconnect network. An optional front end access module provides fan out to OC12/STM4, OC3/STM1, DS3, or E3 interfaces, and an optional expanded interconnect module, sometimes referred to as a hyperconnect fabric, allows dynamic bandwidth expansion of the communication node to include up to eight interconnected local interconnect modules, thereby providing 160 Gbs of essentially non-blocking bandwidth.
Yet another embodiment of the invention enables service providers to offer enterprises differing grades or quality of service (QoS).
Briefly described, an interconnect network according to one embodiment of the invention is incorporated in a communication node having a local communication interface, an associated local interconnect network, and scaling elements. The local communication interface includes a plurality of external communication channels for coupling information into and out of the node and a plurality of internal communication channels for transferring information within the node. Each external communication channel couples to an internal communication channel. The local interconnect network has local transfer elements for directing information between the internal communication channels, and consequently between the external communication channels. The scaling elements enable dynamically scaling the node to include additional local communication interfaces having additional associated local interconnect networks, such that information can be transferred between the local communication interfaces. According to a further feature, as the node expands to include additional local communication interfaces and local interconnect networks, the communication node, optionally, can transfer information between any of the internal communication channels, and thus any of the external communication channels, of the local communication interfaces.
Since an enhanced feature of the invention is dynamic bandwidth scalability, according to further embodiment, the communication node provides an additional local communication interface and an additional local interconnect network; and the scaling elements include an expanded interconnect network. The additional local communication interface has an additional plurality of external communication channels for coupling information in and out of the node, and an additional plurality of internal communication channels for transferring information within the node. The internal and external communication channels of the additional local communication interface couple to each other. The additional local interconnect network includes additional local transfer elements for directing information between the additional plurality of internal communication channels. The local interconnect network and the additional local interconnect network both include non-local transfer elements for directing information between the internal communication channels and the expanded interconnect network. The expanded interconnect network includes expanded transfer elements for directing information between the local interconnect networks, such that information, optionally, can be transferred between any of the internal communication channels of the local communication interfaces.
In another embodiment, the invention includes up to eight local communication interfaces, with associated local interconnect networks. According to the dynamic bandwidth scalability feature of the invention, the expanded interconnect network remains unchanged, regardless of the number of local communication interfaces, and provides the ability to transfer information between the internal communication channels. Such an embodiment provides an ease of bandwidth scalability absent from prior art technology. In a further embodiment, the communication node can be scaled to change the number of local communication interfaces, while the node is operating transferring information. In this way, a communication node, incorporating an interconnect network according to one embodiment of the invention, can more easily meet a service provider's varying bandwidth needs.
As mentioned above, the invention may provide enhanced QoS features. To provide such features, an interconnect network according to one embodiment of the invention can monitor the availability of communication channels. More particularly, the local interconnect network can include a plurality of transceivers for transferring information between the local transfer elements and the internal communication channels. Each transceiver couples to an associated internal communication channel, and has a corresponding availability status indicative of an availability of that communication channel for transferring information. The local interconnect network may also include a plurality of memory storage queues, having associated ones of the transceivers, and including memory for storing information to be transferred by an associated transceiver. The interconnect networks may further include control elements for setting the status corresponding to a particular internal communication channel to indicate unavailability for transferring information, in response to an associated memory queue reaching a selectable content level. In this way, the communication node lowers the likelihood of losing information or blocking transfer due to overloading a particular channel. Additionally, according to a further embodiment, the information coupled into the communication node is assigned a particular priority, and the interconnect networks optionally includes control elements for setting the status corresponding to a particular channel to indicate availability for receiving information having a particular priority, such as high, medium or low, in response to an associated memory queue reaching a selectable content level.
According to a related embodiment, the interconnect networks provides a back pressure signal to the internal communication channels, wherein the back pressure signal contains the availability status for each of the internal communication channels. A further enhancement of this feature utilizes communication bits, initially reserved for a destination address or handle, associated with a particular internal communication channel, to transfer the back pressure / availability status from a local interconnect network to an associated local communication interface. According to another embodiment, the invention provides enhanced error correction. As a result the local interconnect network includes elements for generating a redundant version of information transferred from the local interconnect network to the local communication interface. If error detection elements detect an anomaly in transferred information, error correction elements can recover an error-free version of information from the redundant version. According to a further embodiment, the communication node includes control elements for deactivating those elements, be they line cards or interconnect elements, causing the detected anomalies. According to a related embodiment, the communication node provides improved fault-tolerance by deactivating failed line cards or interconnect elements, without compromising the speed with which information is transferred through the node. Additionally, to provide reduced down-time, the communication node generally, and the interconnects specifically, may also include circuit protection elements for enabling the hot replacement of failed components, while the communication node continues to transfer information.
In a related embodiment, the interconnect networks transfer information internally as information cells, wherein each cell includes groups of information words, and each group of information words is transferred by way of a different internal communication channel. The local interconnect network generates the redundant version by performing a bit-by-bit "exclusive or" operation on pairs of groups of information words, prior to the pair being transferred to the local communication interface. The local interconnect network also transfers the "exclusive or" version of the pair to the local communication interface. In response to a detected anomaly in either member of the pair, the local communication interface can reconstruct an error-free version of the anomalous member by performing an "exclusive or" operation between the non-anomalous member and the "exclusive or" version of the transferred pair. One way to enhance the non-blocking feature of the invention and thus, the speed with which information can be transferred through the interconnect networks, is to avoid the need for re-ordering groups of information words into a complete cell, subsequent to transfer through an interconnect network. According to one embodiment, the invention employs "clumping" to avoid re-ordering and thus, enhance transfer speed. More specifically, the interconnect networks can include elements for "clumping" or combining a plurality of information cells, and for transferring those clumped cells substantially simultaneously. In a further enhancement, the interconnect networks also include elements for appending "dummy" cells to fill in a partial clump prior to the clump being transferred.
In a related embodiment, an interconnect network according to the invention implements the clumping feature by employing storage queues associated with the transceivers. The storage queues intermediately store groups of information words to be transferred. The interconnect networks can further include detection elements for detecting when groups of words of a plurality of information cells to be included in a clump are stored in a queue, and transfer elements for substantially simultaneously transferring the clumped information cells by coupling the groups of words to transceivers. As the communication node expands to include additional local communication interfaces and associated local interconnect networks, it becomes increasingly important for the expanded interconnect network to select an efficient path through which information passes, thereby avoiding unnecessary delays. Accordingly, in a further embodiment, the invention provides a substantially non-blocking feature. According to the non-blocking feature, the expanded interconnect network can include a forwarding array for storing data indicative of an unblocked local path through the expanded interconnect network. The expanded interconnect network can use at least a portion of the destination address of a group of words of an information cell as a pointer into the forwarding array to select an unblocked path. In a further embodiment, the expanded interconnect network employs a plurality of forwarding arrays, each storing data indicative of a segment of an unblocked path through the expanded interconnect network. Further, the expanded interconnect network can use successive portions of the destination address as pointers into each of the forwarding arrays to select each segment of an unblocked path.
According to a further embodiment, the transfer elements of the local interconnect network and the transfer elements of the expanded interconnect network are essentially identical, and therefore, interchangeable. In such an embodiment, the transfer elements can include a mode selection feature for selecting whether the element is to be used in a local mode or in an expanded mode. Such a feature provides substantial cost savings over prior art systems.
In further aspects, the invention includes methods corresponding to the above described apparatus.
Brief Description of the Drawings
The subject matter regarded as the invention is particularly pointed out and distinctly claimed in the concluding portion of the specification. However, the invention, both as to organization and method of practice, together with further objects and advantages thereof, may best be understood by reference to the following illustrative description taken in conjunction with the accompanying drawings in which like numerals refer to like elements, and FIGURE 1 depicts a plurality of communication networks interfacing through a communication node employing interconnection networks according to the invention;
FIGURE 2 is block diagram of a communication node of the type depicted in FIGURE 1 and incorporating interconnect networks according to the invention;
FIGURE 3 is a more detailed interconnection diagram of an exemplary local line card module and local interconnect module of the type depicted in FIGURE 2;
FIGURE 4 is a table illustrative of the interconnections between a typical line card module and an interconnect module of the types depicted in FIGURES 2 and 3;
FIGURE 5 is a more detailed block diagram showing interconnections between an exemplary local interconnect module and an expanded interconnect module of the types depicted in FIGURE 2;
FIGURE 6 is a table detailing the communication channel connections between an exemplary local interconnect module and the expanded interconnect module;
FIGURE 7 is a table further specifying the connection of communication channels between potential local interconnect modules and the expanded interconnect module;
FIGURE 8 is a functional block diagram of a local interconnect logical plane of the type depicted in FIGURE 3 ;
FIGURES 9A and 9B are conceptual illustrations of high- and low-priority queues;
FIGURES 10A and 10B are flow-charts depicting the operation of the queues of
FIGURES 9 A and 9B; and FIGURE 11 is functional block diagram of an expanded interconnect logical plane of the type depicted in FIGURE 5.
Description of Illustrated Embodiments As briefly described above, the invention is directed to an interconnection network in a communication node. Communication nodes are junctions for transferring communication signals between a plurality of sources. As such, communication nodes may be required to interface with systems employing a variety of communication protocols and operating at differing information transfer speeds. Prior art systems typically require maintenance of a plurality of networks, each being capable of interfacing with a particular type of source. In contrast, a communication node, embodying features of an illustrative embodiment of the invention, can process information entering the node at a variety of speeds and formatted pursuant to a plurality of protocols. By way of example, information can enter and leave the communication node at OC48, OC12/STM4, OC3/STM1, DS3 and E3 speeds. Additionally, information can enter and leave the node in IP- or ATM-based formats.
Another important feature of the invention is dynamic bandwidth scalability. A communication node employing interconnection networks according to an illustrative embodiment of the invention, employs a modular design. The modular design enables a service provider to change the number of communication channels by adding or subtracting physical proximately located modules to or from the communication node. According to a preferred embodiment, the modules include a plurality of I/O interfaces coupled to an associated interconnection network. In a further embodiment of the invention, the communication node employs a two-level interconnection network modularity; a local level and an expanded level. More particularly, a plurality of local interconnection network modules, preferably proximately located with respect to each other, couple to an expanded interconnection network, also preferably located proximate to the local interconnection modules. By changing the number of local interconnection network modules that are "plugged-in" to the expanded interconnection module, a service provider can change the bandwidth of the communication node. Moreover, according to a further embodiment, a service provider can connect and unconnect local interconnect modules while the communication node is operating transferring information, thus, providing dynamic bandwidth scalability.
FIGURE 1 illustrates a typical deployment of a communication node 100 employing interconnection networks according to the invention. As shown, the communication node 100 can support, among other systems, SONET rings 10, WWW server farms 20, dial up IP systems 30, IP over CATV/xDSL systems 40, IP over FR networks 50, multi-service ATM networks 60, ATM transports 70, and internet WDM systems 80. Multi-service ATM access 60 can include CBR circuits 61, voice over ATM circuits 62, internet access 63, FR over ATM data 64, and LANs 65. FIGURE 2 shows a block diagram of a communication node 100 incorporating interconnect networks according to an illustrative embodiment of the invention. The communication node 100 employs a two-level interconnect architecture. Accordingly, the node 100 includes one or more local line card modules 102-116, each having first- level associated local interconnect modules 118-132. In the case where the node 100 includes more than one local line card modules, the communication node 100 further includes a second-level expanded interconnect module 134, sometimes referred to as a hyperconnect fabric. The illustrated expanded interconnect module 134 includes three expanded interconnect boards 136-140. The boards are printed circuit boards holding integrated circuits and other components. According to one embodiment, the local line card modules 102-116 are essentially identical to each other. Similarly, the local interconnect modules 118-132 are also essentially identical to each other. Accordingly, for the most part, the following description focuses on an exemplary local line card module 102 and an exemplary local interconnect module 118; that description being equally applicable to the remaining local line card modules 104-116, and the remaining local interconnect modules 120-132.
The local line card module 102 transfers information into and out of the communication node 100, by way of a plurality of I/O interfaces. Those I/O interfaces can be, for example, IP or SONET/SDH ports that accept an OC48 data stream. For grooming to lower-speed interfaces, optional access modules 162-180 can be employed to provide OC12/STM4, OC3/STM1, DS3 and E3 ports. Access module 162 multiplexes input data streams into an OC48/STM16 uplink to local line card module 102. Line card module 102 couples information to an associated local interconnect module 118 by way of a plurality of Gigabit Ethernet connections 142. The local interconnect module 118 transfers information between the I/O channels of local line card module 102.
A feature of the local line card module 102 is that it supports a number of datalink layer encapsulations, implemented by a flexible encapsulation/decapsulation mechanism. The decapsulation mechanism is adaptable to accommodate emerging encapsulations. According to one embodiment, local line card module 102 supports IP over ATM over SONET/SDH; IP over PPP over SONET/SDH; IP over FR over SONET/SDH; IP over PPP over FR over SONET/SDH; IP over PPP over ATM over SONET/SDH; IP over MPLS over SONET/SDH; IP over SNAP 802.2; and IP over Ethernet 2.0. Line card module 102 also supports FRAME Relay DTE.
Those skilled in the art of communication nodes will appreciate that other encapsulations may be accommodated by the invention. The above list is intended to be illustrative, rather than limiting in nature. A further feature of the illustrated communication node 100 is that it is dynamically bandwidth scalable. More particularly, according to one embodiment, the communication node 100 can include up to seven additional local line card modules 104-116, coupled to associated local interconnect modules 120-132 by way of Gigabit Ethernet connections 144-160. The expanded interconnect module 134 transfers information between local interconnect modules 118-132 by way of a plurality of
Gigabit Ethernet connections 170. Each local interconnect module 118-132 is coupled to all three expanded interconnect boards 136-140. Another feature of the invention is that the same expanded interconnect module 134 can be employed for two local line card modules as is employed for additional local line card modules. FIGURE 3 is a more detailed schematic block diagram of an illustrative embodiment 200 of the communication node 100, but having only a single local line card module 102 coupled to a single local interconnect module 118. Since there is only a single local line card module 102, there is no need for the communication node 200 to include the second-level expanded interconnect module 134 of the type shown in FIGURE 1. The local line card module 102 includes eight local line cards 202-216. Local line cards 202-216 are printed circuit boards holding integrated circuits and other components. Each line card 202-216 has six internal (I/O) ports 202a-202f, and an external SONET I/O port 202g. Line card 202 couples information between external I/O port 202g and internal I/O ports 202a-202f. External I/O port 202g couples information into and out of the node 200, and the internal I/O ports 202a-202f connect with up to forty-eight internal communication lines 217 and couple information between the local line card module 102 and the local interconnect module 118. Typically, each internal I/O port a-f includes a Gigabit Ethernet transceiver, providing a Gigabit Ethernet input channel and a Gigabit Ethernet output channel. Preferably, the input and output channels provide 10-bits of information. However, it should be noted that the term transceiver, as used throughout this description, is also intended to encompass structures including separate receivers and transmitters. The external I/O port 202g is preferably software configurable for either SONET or SDH operation. Thus, physical interfaces are software configurable for OC48 or STM16. SONET and SDH PAMS may be freely intermixed within access module 142. A fully loaded local line card module 102 can have up to eight external SONET/SDH I/O ports and forty-eight corresponding internal I/O ports.
The local interconnect module 118 includes three identical interconnect boards 218-222. The interconnect boards 218-222 are printed circuit boards holding integrated circuits and other components. Each board 218-222 is logically subdivided into two essentially identical planes. By way of example, interconnect board 218 includes logical planes 218a and 218b; interconnect board 220 includes logical planes 220a and 220b; and interconnect board 222 includes logical planes 222a and 222b. The communication node 200 transfers information through the interconnect boards 218-222 by way of Application Specific Integrated Circuits (ASICs) 224-228. Each ASIC 224-228 logically includes an a-half and a b-half. The logical a-half services the logical a-plane of a particular interconnect board 218-222, while the logical b-half services the logical b-plane of the particular interconnect board 218-222. By way of example, ASIC 224a services logical plane 218a and ASIC 224b services logical plane 218b. The illustrated embodiment of FIGURES 2 and 3 is described with respect to particular physical groupings of line cards 202-216 and local interconnect boards 218- 222, and particular logical divisions of local interconnect planes 218a, 218b, 220a, 220b, 222a and 222c. However, as one skilled in the art will appreciate, the number of line card modules 102-116 and associated local interconnect modules 118-132 can be varied. Moreover, the number of local line cards included in a local line card module 102-116 can also be varied. Further, the number of local interconnect boards included in a local interconnect module 118 and the number of extended interconnect boards included in an extended interconnect module 134 can be varied. All logical divisions can also be varied.
With that caveat, according to the illustrated embodiment, each local interconnect board 218-222 includes sixteen internal communication ports (eight associated with each logical plane), and sixteen expanded communication ports (eight associated with each logical plane). As each interconnect board is essentially identical, board 218 will be discussed in detail. Boards 220 and 222 have a similar construction and operation. Specifically, local interconnect board 218 has eight internal communication ports 0a-7a, associated with ASIC 224a, and eight internal communication ports 0b-7b, associated with ASIC 224b. Local interconnect board 218 also includes eight expanded communication ports 8a- 15a, associated with ASIC 224a, and eight expanded communication ports 8b- 15b, associated with ASIC 224b. Each internal and expanded communication port includes an Ethernet transceiver providing a Gigabit Ethernet input channel and a Gigabit Ethernet output channel. Each internal communication port 0a-7a and 0b-7b couples to an internal communication port a-f of a line card 202-216, and transfers information between the local line card module 102 and the local interconnect module 118. Similarly, each internal communication port a-f of line cards 202-216 couples to an internal communication port 0a-7a and 0b-7b of one of the interconnect boards 218-222.
FIGURE 4 shows a table 300 illustrative of the interconnections between the internal communication ports of line cards 202-216 and the internal communication ports of interconnect boards 218-222. Column 302 shows the line card designation numbers 202-216. Column 304 lists sets of ordered pairs in the form (interconnect plane #, port #), wherein those ordered pairs identify which internal port 0a-7a and 0b-7b of interconnect module 118 couples to each internal port a-f of line cards 202-216. By way of example, the first line of table 300 indicates that line card 202, internal port a, couples to internal port 0a of interconnect plane 218a. Similarly, internal ports b, c, d, e, and f of line card 202, couple to internal ports 0b of plane 218b, 0a of plane 220a, 0a of plane 220b, 0a of plane 222a, and 0b of plane 222b, respectively. In other words, the internal communication ports of line card 202 couple to the "zero numbered" ports of the interconnect planes 224a, 224b, 226a, 226b, 228a and 228b. In a similar fashion, the internal communication ports of line card 204 couple to the "one numbered" ports (e.g. la). Those of line card 206 couple to the "two numbered" ports (e.g. 2a). Those of line card 208 couple to the "three numbered" ports (e.g. 3a), and so on.
According to an illustrative embodiment, and as shown below in TABLE 1 , the communication node 200 transfers information from the local interconnect module 118 to associated local line card module 102 in 64-byte cells.
As shown in TABLE 1, the 64-byte cell is subdivided into 16-byte groups. Logical plane 218a transfers 8-words of 2-bytes each. Logical planes 218b, 220a and 220b do the same. The two least significant bytes (LSBs) of the first and third 16-byte groups (i.e., the groups transferred by logical planes 218a and 220a) are used for the address/handle of a destination line card. The two LSBs of the second and fourth 16- byte groups (i.e., the groups transferred by logical planes 218b and 220b) are used for interconnect addressing and flow control information. Board 222 provides error correction and redundancy information. More particularly, logical plane 222a provides a bit-by-bit "exclusive or" (θ) between the information transferred on logical plane 218a and logical plane 220a. Logical plane 222b provides a bit-by -bit "exclusive or" between the information transferred on logical plane 218b and logical plane 220b. In the illustrated embodiment of TABLE 1 , a byte contains 8-bits and a word contains 2-bytes. However, those skilled in the art will appreciate that alternative byte and word conventions may be employed. TABLE 1
Figure imgf000018_0001
With each line card 202-216 having six Gigabit internal Ethernet ports a-f, spread across three interconnect boards 218-222, and according to the format of TABLE 1, the six Gigabit internal Ethernet ports a-f provide 3 -Gbs of usable bandwidth. More specifically, board 222 is not used for payload bandwidth, instead providing redundancy and error correction information, thus leaving 4-Gbs of bandwidth. 4-bytes out of sixty- four contained in a transferred cell (the LSBs of logical planes 218b and 220b) are used for interconnect addressing and flow control information, leaving 3.5 Gbs of bandwidth. And, 4-bytes out of the remaining fifty-six (the LSBs of logical planes 218a and 220a) are used by the line cards 202-216 as a destination handle/address, leaving 3 Gbs of bandwidth. This ensures that the communication node 200 can provide a sustained OC- 48 (2.4 Gbs) transfer rate TABLE 2 below depicts a typical information cell format for information transferred from a line card 202-216 to local interconnect planes 218a, 218b, 220a, 220b, 222a and 222b.
As shown in the first column of TABLE 2 and as previously described with respect to TABLE 1, information is transferred in 8-word/16-byte groups. Each logical plane 218a, 218b, 220a and 220b receives a 16-byte group. Logical plane 222a receives the "exclusive or" of planes 218a and 220a, and logical plane 222b receives the "exclusive or" of logical planes 218b and 220b. Bytes 2-16 of logical planes 218a, 218b, 220a and 220b provide the transferred data.
TABLE 2
Figure imgf000019_0001
"X" above represents an XOFF from the line card to the local interconnect 118. Information cells pass through the local interconnect 118 on separate planes 218a-222b, but with some discrepancies. By way of example, the a-planes contain the line card destination addresses. The b-planes allow for 14-bits of extra "payload" data carried through untouched. The payload byte above typically has its most significant bit (bit "P") set as parity for words 0 and 1 together. Even on plane 222a, the "P" bit covers the 15-bits which precede it, rather than the parity across planes 218a and 220a.
The line card destination address is an address or handle (global to the node 100 system wide) which specifies the destination line card (for unicast information) or line card set (for multicast information) to which the information cell is to be transferred.
TABLE 3 below depicts a preferred destination address format for locally transferred unicast information, while TABLE 4 depicts a preferred destination address format for locally transferred multicast information.
TABLE 3
Figure imgf000019_0002
TABLE 4
Figure imgf000020_0001
Byte-0 of TABLE 3 provides an "APS" bit, a Line Card Module designation field and a Line Card designation field. As shown in FIGURE 3, according to the illustrated embodiment, there are eight potential destination line cards 202-216. Each one is assigned a 3-bit code 000-111. Bits 3, 4 and 5 of byte-0 of TABLE 3 provide this code. As shown in FIGURE 2, a communication node 100, according to the illustrated embodiment, has eight potential line card modules 102-116. Each line card module 102- 116 is assigned a 3-bit code 000-111. Bits 0, 1 and 2 of byte-0 of TABLE 3 provide this code. Bytes 2-15 provide the transferred data. The "P" bit provides parity across byte 1.
According to a preferred embodiment, unicast information cells travelling to a single destination line card have the "Pri" bit of TABLE 3 set to one for high-priority traffic. If the "APS" bit of TABLE 1 is set, the cell is sent to both the designated line card (n) and the (n+1) line card.
With reference to TABLE 4, the multicastID is an address into a 16 k x 9-bit RAM 748 of FIGURE 8. This 14-bit address retrieves a 9-bit value. The least significant 8-bits are a mask with each bit representing a line card 202-216. The most significant bit represents priority; set to a logical one for high priority or a logical zero for low priority. If the node includes more than one local line card chassis 102, then each bit of the 8-bit mask represents a particular local line card chassis.
In operation, and as illustrated in TABLE 1 above, the communication node 200 transfers each 16-byte group over a different internal communication channel. By way of example and referring again to FIGURE 3, assume each line card 202-216 has an associated address, and information enters line card 202 by way of external port 202g. Assume further that the entering information has a destination address of line card 208. As shown in FIGURE 2, internal port 202a couples a first 16-byte group to internal port 0a of interconnect board 218. Internal port 202b couples a second 16-byte group to internal port 0b of board 218. Internal port 202c couples a third 16-byte group to internal port Oa of board 220, and internal 202d couples a fourth 16-byte group to internal port Ob of board 220. Internal port 0a of board 218 couples the first 16-byte group to ASIC 224a. Internal port Ob of board 218 couples the second 16-byte group to ASIC 224b. Internal port 0a of board 220 couples the third 16-byte group to ASIC 226a, and internal port Ob of board 220 couples the fourth 16-byte group to ASIC 226b.
ASIC 224a processes the line card destination address and directs the first 16- byte group to internal port 3a of board 218. ASIC 224b processes the line card destination address and directs the second 16-byte group to internal port 3b of board 218. ASIC 226a processes the line card destination address and directs third 16-byte group to internal port 3a of board 220. ASIC 226b processes the line card destination address and directs the fourth 16-byte group to internal port 3b of board 220. Board 222 generates a bit-by-bit "exclusive or" between the first and third groups, and between the second and fourth groups. ASIC 228a processes the line card destination address and directs the "exclusive or" combination of the first and third groups to internal communication port 3a of board 222, and ASIC 228b processes the line card destination address and directs the "exclusive or" combination of the second and fourth groups to the internal communication port 3b of board 222. Board 222 in turn couples the first, second, third and fourth groups to destination line card 208, internal ports 208a-208d, respectively. Similarly, board 222 couples the "exclusive or" version of the groups of bytes to internal ports 208e and 208f. Destination line card 208 then performs various types of error checking, such as plain parity, 8B10B disparity and CRC across multiple cells. If line card 208 detects a bad character error, software can alert the line card to use the "exclusive or" version to retrieve an error-free version of the transferred information. According to a further embodiment, the line card module 102 can determine whether a detected error is due to a failed line card 202-216 or a failed interconnect board 218-222. In the case where the error is due to a failed line card, the interconnect module removes that card from operation. In the case where the detected error is due to a failed interconnect card 218-222, the line card detecting the error can signal the error prone interconnect board 218-222 to take itself off line. So as not to compromise bandwidth, interconnect board 222 can automatically take the place of either interconnect board 218 or 220, until the failed board is replaced. According to a further feature, the failed board can be hot-swapped.
As discussed above, a feature of the invention is that according to a preferred embodiment, the communication node 100 is dynamically bandwidth scalable to include additional line card modules 104-116, having additional associated local interconnect chassis 120-160. According to a preferred embodiment, the modular construction of the line card modules 102-116, along with the modular construction of the local interconnect modules 118-132, in combination with the expanded interconnect module 134 provides the scalable feature. More specifically, as indicated in FIGURE 2, according to the illustrated embodiment 100, a first mechanical chassis 101 houses up to eight local line card modules 102-116. Similarly, a second mechanical chassis 103 houses up to eight local interconnect modules 118-132. Further, a third mechanical chassis 105 houses the extended interconnect network 134. The first mechanical chassis 101 electrically couples to the second mechanical chassis 103 by way of communication channels 142- 160. The second mechanical chassis 103 electrically couples to the third mechanical module 105 by way of communication channels 170. The communication node 100 employs connectors designed for "hot-swapping" at module interfaces. Those connectors enable the local line card modules 102-116, the local interconnect modules 118-132, and the expanded interconnect boards 136-140 to be connected and unconnected (i.e., "hot-swapped") from their respective mechanical chassis 101, 103 and 105, while the communication node 100 is powered and operating transferring information. Thus, as a service provider requires additional bandwidth, additional local line card modules 102-116, with associate additional local interconnect modules 118- 132 can be "plugged in" to chassis 101 and 103, respectively. FIGURE 5 is a more detailed schematic block diagram 400 showing interconnections between an exemplary local interconnect module 118 and an expanded interconnect module 134. The embodiment of FIGURE 4 is illustrative of the dynamic bandwidth scalability of the communication node 100. As mentioned above, the local interconnect module 118 is essentially identical to optional local interconnect modules 120-132. Accordingly, the following illustrative description focuses on the interconnections between local interconnect module 118 and expanded interconnect module 134. However, each additional local interconnect module 120-132 interfaces to expanded interconnect module 134 in a like manner.
As shown in FIGURE 5, and as discussed in more detail with regard to FIGURE 3, the local interconnect module 118 includes three interconnect boards 218-222. Each board 218-222 logically subdivides into an a-plane and a b-plane. The a-plane and the b-plane are essentially identical and share an ASIC, which is also logically subdivided into an a-half and a b-half. Each ASIC 224-222 provides, among other functions, routing and switching between the internal communication channels associated with the particular interconnect board 218-222. According to the illustrated embodiment, each ASIC 224-226 may also provides storage queues, a control processor, control registers and status registers. It also controls Gigabit Ethernet I/O interfaces included on each interconnect board 218-222. The ASICs 224-228 further provide one or more pointer queues for storing information indicative of a substantially non-blocked path through its switching/routing circuitry. Each local interconnect board 218-222 includes sixteen internal I/O ports 0a-7a and 0b-7b. The internal I/O ports 0a-7a and 0b-7b provide Gigabit Ethernet interfaces. As shown in FIGURE 3, those Gigabit interfaces couple to the internal communication channels of an associated local line card module 102. Each local interconnect board 218-222 also provides sixteen expanded I/O ports 8a- 15a and 8b- 15b. Each expanded I/O port 8a- 15a and 8b- 15b provides a Gigabit input channel and a Gigabit output channel. The expanded I/O ports 8a-15a and 8b-15b couple information between the local interconnect module 118 and the expanded interconnect module 134.
The expanded interconnect module 134 includes three essentially identical expanded interconnect boards 136-140. Each board 136-140 includes, among other components, one hundred and twenty-eight Gigabit Ethernet transceivers. Each board 136-140 also includes four ASICs 402-408, 410-416 and 418-424, respectively. ASICs 402-424 are essentially identical to ASICs 224-228. However, ASICs 404-424 are mode selected to operate in an expanded interconnect mode, rather than the local interconnect mode of ASICs 224-228. As in the case of ASICs 224-228, ASICS 404-424 each logically subdivides into an a-half and a b-half. Each half includes sixteen Gigabit Ethernet I/O ports, wherein each port includes a Gigabit input channel and a Gigabit output channel. Each of the sixteen Gigabit Ethernet ports couple to a Gigabit transceiver on the extended interconnect board.
By way of a specific example, board 136 of FIGURE 5 includes ASICs 402-408. ASIC 402 is subdivided into to logical halves 402a and 402b. Similarly, ASIC 404 is subdivided into logical halves 404a and 404b; ASIC 406 is subdivided into logical halves 406a and 406b; and ASIC 408 is subdivided into logical halves 408a and 408b. ASIC 402 includes Gigabit Ethernet ports 0a-15a, on half 402a, and 0b- 15b on half 402b. Ports 0a-15a couple to transceivers 0-15 on board 136, and ports 0b-15b couple to transceivers 16-31. Gigabit ports 0a-15a and 0b- 15b of ASICS 404-408 successively couple to remaining transceivers 32-127. Gigabit I/O ports of ASICs 410-416 and 418- 424 couple to one hundred and twenty-eight transceivers of boards 138 and 140, respectively, in an identical fashion to that described with respect to ASICS 402-404 on board 136.
FIGURE 6 is a table 500 illustrating the communication paths between local interconnect module 118 and expanded interconnect module 134. Column 502 specifies the local interconnect port designation. Column 504 specifies the communication paths between local interconnect board 218 and expanded interconnect 134. Similarly, column 506 specifies the communication paths between local interconnect board 220 and expanded interconnect 134; and column 508 specifies the communication paths between local interconnect board 222 and expanded interconnect 134. As can be seen from FIGURES 4 and 5, the communication channels of board 218 couple to board 136. Those of board 220 couple to board 138, and those of board 222 couple to board 140. As also can be seen from FIGURE 5, the I/O interfaces from local interconnect module 118 couple to the zero and eight ports of ASICs 402-424. FIGURE 7 is a table 600 depicting the expanded interconnect port assignments for each of the potential local interconnect modules 118-132 of FIGURE 1. Column 602 identifies the particular local interconnect modules 118-132. Column 604 identifies the port assignments for the extended interconnect module 134. As shown, local interconnect module 120 couples to the one and nine ports; module 122 couples to the two and ten ports; module 124 couples to the three and eleven ports; module 126 couples to the four and twelve ports; module 128 couples to the five and thirteen ports; module 130 couples to the six and fourteen ports; and module 132 couples to the seven and fifteen ports.
TABLE 5 below specifies a preferred format for the destination address for unicast information transferred from a local interconnect modules 118-132 to the expanded interconnect module 134. Similarly, TABLE 6 specifies a preferred format for the destination address for multicast information transferred from a local interconnect modules 118-132 to the expanded interconnect module 134.
TABLE 5
Figure imgf000025_0001
Referring to TABLE 5, bit 7 of the most significant byte (MSB) is the parity bit, which represents parity across the previous 15-bits of the destination address. Bit 5 is the "Valid" bit. The "Valid" bit is set if the destination address is valid. Bit 4 is the "Clump" bit. The "Clump" bit is set if there is a valid combination or clump of cells. Clumping is a feature of the invention employed for eliminating the need for reordering transferred information subsequent to transfer. As discussed in further detail below, with reference to FIGURE 7. Bit 7 is the "Pri" bit. The "Pri" bit is set to one for high- priority traffic. Bit 7 of the LSB is the "APS" bit. The "APS" bit is simply passed through the expanded interconnect 134 to local interconnects 118-132. Bits 3-5 of the LSB provide the 3-bit designation 000-111 for the destination line card module. Bits 0-3 of the LSB provide the 3-bit designation 000-111 for the particular line card.
TABLE 6
DestHi DestLo
7 6 5 4 3 2 1 0 7 6 5 4 3 2 1 0
P 1 Mu ticastID (not 0 x 3FFF Preferably, the "Multicast ID" of TABLE 6 is passed on to the expanded interconnect module 134 to be translated. If the payload data portion of the cell is 0x3 FFF, the information cell is considered to be invalid.
TABLES 7 and 8 depict a preferred format for unicast and multicast destination addresses, respectively, for information cells transferred from the expanded interconnect module 134 to a local interconnect modules 118-132. As described above, the "P" bit provides parity across the destination address. The "Valid" bit is set if the destination address is valid. The "Pri" bit is set for high-priority traffic. The "APS" bit is set if the cell is to be sent to both the designated line card (n), and the (n+1) line card. Bits 0-3 provide the designation code for the line card to which the cell is sent.
TABLE 7
Figure imgf000026_0001
TABLE 8
Figure imgf000026_0002
FIGURE 8 is a functional block diagram of an illustrative local interconnect board of FIGURE 5. The board 218 includes an ASIC 224 of the type employed in the local interconnect module 118. Preferably, ASIC 224 is identical to ASICs 402-424 employed in the expanded interconnect module 134. As all of the local interconnect boards are preferably identical, for the purpose of the following discussion it is assumed that board 218 is an exemplary local interconnect board of local interconnect module 118, and interfaces with local line card module 102 of FIGURE 2. However, those of ordinary skill in the art will appreciate that the ASIC 224 need not be identical to ASICs 404-424. As shown in FIGURE 8, the interconnect board 218 includes Gigabit transceiver sets 704 and 708, memory elements 710, controller 712 and status and control registers 714. Gigabit transceiver set 708 provides Gigabit I/O ports 0a-7a and 0b-7b, which couple to the internal communication channels of a local line card module 102 of FIGURE 3. Gigabit transceiver set 704 provides Gigabit I/O ports 8a- 15a and 8b- 15b, which couple to the expanded communication channels of the expanded interconnect module 134, shown in FIGURE 5.
Each transceiver of sets 704 and 708 couples to the ASIC 224 by way of associated input and output shift and hold registers. More specifically, transceivers of set 704 couple to input shift and hold registers 714 by way of lines 716 and output shift and hold registers 718 by way of lines 720. Transceivers of set 704 couple to input shift and hold registers 722 by way of lines 724, and output shift and hold registers 726 by way of lines 728.
The ASIC 224 also includes a dual-port RAM 730 for storing various stacks and queues 731 associated with flow control information. Flow status 733 stores an availability status, regarding the availability of a particular line card to receive information. RAM 730 intermediately stores information being transferred through the board 218. Shift and hold registers 714 and 716 couple to the dual-port RAM 730 by way of lines 732 and 734, respectively. Shift and hold registers 722 and 726 couple to the dual-port RAM 730 by way of lines 736 and 738. The dual-port RAM 730 also couples to destination stack 740 by way of lines 742. The ninety-six destination queues 740 intermediately store addresses representative of where particular data is stored in RAM 730. The queues 740, preferably employ a plurality of stacks for ease of addressing. However, other storage structures can be employed. As discussed above in the Summary of the Invention, and as discussed in further detail below, according to a preferred embodiment, the invention employs a plurality of memory storage queues / buffers to aid in the efficient transfer of information. It should be noted that the terms queue and buffer are used interchangeably. The dual-port RAM 730 provides an output queue for each transceiver of sets 704 and 706. More specifically, information cells coupled into board 218 to be transferred to a line card 202-204 of local interconnect 102, are first written into buffer memory at an address which is written into an output queue. Free list memory 742 provides a list of available buffer memory addresses. There is a reference counter 744 for each of the 1536 buffers in the dual port RAM 730. Reference counter 744 contains the number of output queues to which the contents of the respective buffers are to be sent. A reference counter 744 decrements in response to information being read from an associated buffer. When the reference counter reaches zero, the address of the buffer is returned to free list 743. In this way, the ASIC 224 can track the available buffer locations associated with each transceiver. Information written to buffer memory is subsequently transferred to one of the output shift and hold registers 720 or 728, and held there until an internal time slot arrives in which the destination address lookup can be performed, the read from the free list memory 742 can be performed, the write to the buffer memory can be performed, and the write to the output queue can be performed.
According to a preferred embodiment, the invention provides enhanced QoS features. To that end, queues 731 can include QoS queues. The QoS queues, such as those conceptually illustrated in FIGURES 9A and 9B, can have multiple watermark levels; those levels corresponding to differing priorities. By way of example, high- priority queue 900 of FIGURE 9A can have two watermarks 906 and 908. In range 902, queue 900 reports its status as "stop-none," indicating the I/O channel associated with queue 900 is ready to receive information of any priority. As depicted in the flow-chart of FIGURE 10A, during operation, in range 904, queue 900 reports its status as "stop- low" 1002, indicating the I/O channel associated with queue 900 is ready to receive information having a "medium" priority or higher, as indicated at 1004. When the queue 900 is filled up to level 906, it reports its status as "stop-all," as shown at 1006. As shown at 1008, this indicates that its associated I/O channel is unavailable. If the "Stop-Low" watermark 908 of queue 900 has not been reached, as indicated at 1006, it is available to receive information of any priority.
Low-priority queues, such as queue 910 depicted in FIGURE 9B, the operation of which is illustrated in flow-chart 1010 of FIGURE 10A, can include three watermarks 918, 920 and 922. As shown at 1012, queue 910 reports a status of "Stop-None" in range 912. As shown at 1014 and 101B, it reports a status of "Stop-Low" in range 914. As shown at 1018 and 1020, it reports a status of "Stop-Medium" in range 916, and as shown at 1022 and 1024, queue 910 reports a status of "Stop- All" subsequent to reaching watermark 918. High-priority queues, such as queue 900, enable associated line cards to pass low- and medium-priority traffic, while not allowing low-priority traffic of one line card to strangle medium-priority traffic of a different line card.
Low-priority queues, such as queue 910, enable associated line cards to pass low-, medium- and high-priority traffic, while not allowing low-priority and medium- priority traffic of one line card to strangle high-priority traffic of a different line card. It also prevents low-priority traffic of one line card from strangling medium- and high- priority traffic of a different line card.
To efficiently manage information of differing priorities, the dual-port RAM 730 preferably provides storage for sixty-four low-priority unicast queues; one for each possible local line card in the communication node 100. The RAM 730 also provides storage for sixteen high-priority unicast queues; one for each line card of its local interconnect module, one for each potential additional local interconnect module, and one extra queue. Multicast traffic, preferably employs four low-priority and four high- priority queues.
Additionally, each plane of the expanded interconnect 134 employs eight high- priority unicast queues; one for each potential local interconnect module 118-132. Each expanded interconnect logical plane also employs eight high-priority and eight low- priority multicast queues; again, one for each potential local interconnect destination module 118-132.
A related component, the queue depth logic circuitry 746, maintains a status of all of the line cards 202-216 of local module 102. The status provides information regarding the availability of each line card 202-216 to receive information of varying priority levels. Another feature of the illustrated embodiment of the invention is the way in which the node 100 passes the flow control status (sometimes referred to as back pressure status) between the expanded interconnect module 134 and each of the line cards of the local interconnect modules 118-132. According to one preferred embodiment, the invention utilizes bits of the information cell, previously reserved for the destination address. These bits are indicated in TABLE 1 as the "Flow Control" words on the b-channels. Flow control information is passed between the local interconnect modules 118- 132 and the expanded interconnect module 134 using the least significant word of the b- channel. These bits are included in the parity calculation of the parity bit in the primary channel's destination address word. This format is generally illustrated above in TABLE 1 , with respect to local interconnect plane 218a word 0, and local interconnect plane 218b word 0. This flow information is preferably not repeated on all links. As illustrated in TABLE 9 below, also with reference to local interconnect planes 218a and 218b and expanded interconnect board 136, flow control information is sent in a two- cell sequence. More particularly, column 1 of TABLE 9 lists the expanded interconnect port in
(ASIC reference designation, ASIC port designation) format. Column 2 lists the port reference designations for local interconnect plane 218b. Type 0 and Type 1 identifies the information contained in the byte (e.g. if local interconnect port 8b receives a Type 0 byte, that byte contains Low6, Low5, Low4, Low3, Low2, Lowl and Low 0 flow control information). Each of the Low0-Low63 bits are set if the corresponding low- priority queue is not full and thus, can receive data. Similarly, the High0-High7, MCHigh, and MCLow bids are set if the corresponding high-priority, multicast high- priority and multicast low-priority queues have space available for receiving information.
TABLE 9
Figure imgf000031_0001
Even though the high-priority, and some of the low-priority flow-control information is repeated on both cycles, there may nevertheless be some associated latency. Thus, to avoid queue overflow, the watermark levels are programmed at a level that takes into account potential latency. By way of example, if flow control latency takes four cells to stop incoming information, (4* 16)-4 locations should be reserved above the watermark to avoid overflow. This results from each of sixteen local input ports potentially aiming at the queue for four cell times. Thus, it would be draining out four information cells in that interval. In contrast, the space below the watermark level need only be (1 *flow control latency) to avoid underflow. As a result, a preferred embodiment sets the watermark threshold levels between twelve and eighteen bytes out of one hundred and ninety-two bytes. According to the above-discussed structures and protocols, the interconnect networks support Constant Bit Rate (CBR), Variable Bit Rate-Real-Time (VBR-rt), Variable Bit Rate-Non-Real-Time (VBR-nrt), and Unspecified Bit Rate (UBR) QoS categories. The interconnect networks can operate as a class-based ATM switch. Thus, traffic is queued for transfer based on the service category of the virtual circuit. However, shaping and policing are performed on a per-virtual-circuit basis. The interconnect networks also support QoS features for IP networks, such as the Differentiated Services Model.
As also mentioned above, a preferred embodiment of the invention employs "clumping" to increase the rate with which information can be transferred through the interconnect networks. Typically, in prior art systems, portions of communications can pass through an interconnect network at varying speeds, thus arriving at a common destination in a misordered fashion. Reordering information subsequent to transfer can waste valuable time, and has the potential for receiving out of order cells. Therefore, according to a preferred embodiment of the invention, the expanded interconnect network 134 includes elements for "clumping" or combining a plurality of information cells and for transferring the clumped cells substantially simultaneously.
More particularly, the queue depth logic 746 detects when a group of four unicast information cells are available in a single queue. In response to detecting four unicast cells in a single queue, the queue depth logic 746 signals the dual-port RAM 731 working in conjunction with the destination stack 740 to transfer the detected four cell clump to shift and hold registers 726 for substantially simultaneous transfer via Ethernet transceivers 708. With the clump of cells being transferred together, they arrive at a destination within a close enough time proximity to avoid reordering. According to a further embodiment, a programmable wait timer begins decrementing upon the arrival of a first information cell to be included in the clump. If the timer expires prior to the complete formation of a clump, it triggers the cell(s) ready to be sent to be combined with 4-N invalid cells, where N is the number of cells which the clump is lacking. Multicast cells are clumped together across paths. When multicast traffic is available to be sent on at least four different paths, it is considered available for transmission. As in the case of unicast traffic, a programmable wait timer on any given multicast queue can artificially render multicast traffic eligible. A programmable watermark threshold on multicast queues can also artificially render multicast traffic eligible. Whenever multicast traffic is eligible to be sent, "Qlnfo" cells are sent on the remaining links to or from the expanded interconnect 134. The ASIC 224 also includes a translation memory 748. The translation memory
748 provides storage for path segments through the expanded interconnect module 118, if such a module is included in the system. In the case of a node configured as shown in FIGURE 2, translation memory 748 preferably contains nine logical storage areas; one for each local interconnect modules 102-116, and one for the expanded interconnect module 118. The expanded interconnect storage area is configured as a bitmap of destination line cards and priority. Destination address circuitry 750 accesses the translation memory 748, and the multicast bitmap register 752 receives the accessed information.
A feature of the invention is synchronization of the local interconnect boards of a particular local interconnect module with each other, and the synchronization of the expanded interconnect boards 138-140 with each other. Since each board is independent, although they derive their clock frequencies from the same source, some signals are employed to establish and maintain synchronization between boards. Slot synchronization forces essentially identical "time zero" references between boards, and thus planes, in a module. Cell synchronization enables a local interconnect module to set its slot zero reference such that its transmitted information cells can arrive at the expanded interconnect 134 at a safe time. CPU synchronization enables certain CPU write operations to take place at the same time on all planes in a module.
Plane synchronization logic 752 provide signals 754 necessary to synchronize certain read and write operations between each plane of local interconnect module 118. Control registers 758 provide flow control information by way of lines 760 to the other planes of local interconnect 118, if operating in local mode, or the additional expanded interconnect planes of module 134 if operating in expanded mode.
Plane to plane cell synchronization is attained by cellok inter-plane connections 761. An asserted cellok signal 761 indicates that the corresponding plane has received a valid and error free cell header containing the 2-byte destination address. According to the illustrated embodiment, each plane outputs 16 cellok signals 761 and inputs 32 cellok signals 761. Each cellok output, N, represents that both the a-ports and the b- ports have valid cell headers.
For a cell to be forwarded, all operating planes assert their respective cellok signals 761. If one plane asserts cellok signals 761 and other planes do not, errors are recorded in CPU addressable registers 758. If a plane fails, the system has the capability of instructing the operating planes to ignore the failed plane. In this way, a single failed plain does not reduce the rate with which the effected local or expanded interconnect can transfer information.
Substantially identical ASICs are employed in the local interconnect modules 118-132 and the expanded interconnect module 134. To that end, ASIC 224 includes mode select 756 for selecting whether ASIC 224 is to operate as a local interconnection circuit or as an expanded interconnection circuit. As shown in FIGURES 3 and 5, in local mode, ports 0a-7a and 0b-7b connect to local line cards and ports 8a-15a and 8b- 15b connect to expanded interconnect 134. Alternatively, in expanded mode all ports 0a-15a and Ob- 15b connect to local interconnect planes, such as 218-222 shown in FIGURE 5.
Referring again to FIGURE 8, an important feature of the illustrated ASIC 224 is a "slot counter" contained in timers, counters, control registers 758. The slot counter repeatedly counts from 0-15. Each port 01 -15a and Ob- 15b is assigned a slot count. Each time the slot count 0-15 matches a port number, a check is performed to determine if there is a cell to be transmitted out that port. If there is, the cell is copied from RAM 730 to shift and hold register 718 or 726 for transmission. If there is no cell to be transmitted, then a flow control cell is transmitted. According to the illustrated embodiment, a common slot counter is employed for the a-ports and the b-ports. As mentioned above, board 218 also includes controller 712 and memory 710.
Memory 710 stores the control code for board 218. As such, it provides start up initialization of statuses, pointers and communication interfaces. Controller 712 provides a variety of conventional processor functions.
As in the case of the local interconnect boards, expanded interconnect boards divide logically into essentially identical a- and p-planes. Thus, for illustrative purposes, FIGURE 11 is described in terms of the a-plane 136a of expanded interconnect board 136 of FIGURE 5. Plane 136a includes four ASICs 402a, 404a, 406a and 408a. ASICs 402a-408a are essentially identical to ASIC 224 of FIGURE 8. Each ASIC 402a-408a controls sixteen Gigabit Ethernet ports 1102-1108. Ports 1102-1108 couple to ASICs 402a-408a by way of shift and hold registers, such as registers 714 and 718 of FIGURE 8. Bus 110 couples interplane control signals, such as those generated by counters, timers and registers 758 and plane synchronization logic 753, to each ASIC 402a-408a. Processor 1114 controls ASICS 402a-408a by way of bus 1112. Processor 1114 includes a CPU module, DRAM, FPGA control and Ethernet control, much in the same way that memory 710, controller 712 and control and status registers 753 provide these functions for local interconnect board 218.
It should be noted that connections and circuit divisions referred to in the above description may be representative of both actual and logical connections or divisions.
FIGURE 11 shows a functional block diagram of a typical expanded interconnect logical plane illustrative of any of interconnect planes of FIGURE 4. It will thus be seen that the invention efficiently attains the objects set forth above, including providing dynamically bandwidth scalable interconnect network. Since certain changes may be made in the above constructions and the described methods without departing from the scope of the invention, it is intended that all matter contained in the above description or shown in the accompanying drawings be interpreted as illustrative and not in a limiting sense.
Having described the invention, what is claimed as new and protected by Letters Patent is:

Claims

Claims
1. An interconnect network for operation within a communication node, said network comprising, a selectable number of local interconnect modules located proximate to each other and each having local transfer elements for transferring information between a plurality of local I/O channels and for transferring information between said plurality of local I/O channels and a plurality of non-local I/O channels, and an expanded interconnect module located proximate to said local interconnect modules and having coupling means for electrically coupling to said non-local I/O channels, and expanded transfer elements for transferring information between said local interconnect modules.
2. An interconnect network according to claim 1 wherein said local transfer elements include means for synchronizing information transferred between each of said local transfer elements.
3. An interconnect network according to claim 1 wherein said expanded transfer elements include means for synchronizing information transferred between said local interconnect modules.
4. An interconnect network according to claim 1 further comprising hot-swap means for changing the selected number of said local communication modules included in said interconnect network, while said interconnect network is operating transferring information.
5. An interconnect network according to claim 1, wherein said local I/O channels have an associated priority and said interconnect network further comprises QoS means for transferring information from one of said local I/O channels having a relatively higher priority in preference to transferring information from one of said local I/O channels having a relatively lower priority.
6. An interconnect network according to claim 1 , wherein said local I/O channels have an associated availability for receiving information, and said interconnect network further comprises, status means for maintaining a status indicative of said associated availability for one or more of said local I/O channels, and back pressure means for communicating said status out of said local I/O channels.
7. An interconnect network according to claim 6, wherein said local interconnect modules further include a plurality of memory queues, said plurality of local I/O channels having associated memory queues, and said memory queues having means for storing information received by way of an associated local I/O channel.
8. An interconnect network according to claim 7 further comprising means for setting said status corresponding to a particular one of said local I/O channels to indicate unavailability for receiving information in response to said associated queue of said particular local I/O channel reaching a selectable content level.
9. An interconnect network according to claim 7, wherein information coupled into and out of said network has an associated priority, and said network further comprises means for setting said status corresponding to a particular one of said local I/O channels to indicate unavailability for receiving information having a particular priority, in response to said associated queue of said particular transceiver reaching a selectable content level.
10. An interconnect network according to claim 6, wherein information transferred through said interconnect modules includes a data portion and a destination address portion, said destination address portion specifying through which one of said local I/O channels said information is to be transferred out of said interconnect network, and said back pressure means is adapted for replacing said destination address portion with said status for communicating said status out of said interconnect network.
11. An interconnect network according to claim 1 further comprising redundancy generating means for generating an alternative version of information being transferred out of said interconnect network through said local I/O channels.
12. An interconnect network according to claim 11 , wherein said interconnect network is adapted for transferring information as information cells, each of said cells including groups of information words, and wherein said redundancy generating means is adapted for causing said alternative version of said information to be a bit-by-bit "exclusive-or" between pairs of said groups of information words included in an information cell.
13. An interconnect network according to claim 1, wherein said interconnect network is adapted for transferring information as information cells, and said local transfer elements further include means for substantially simultaneously transferring a plurality of said information cells out of said local interconnect modules.
14. An interconnect network according to claim 13 further comprising a memory queue for intermediately storing said plurality of information cells to be transferred, and queue detection means for detecting when a selected number of said information cells are stored in said memory queue, wherein said local transfer elements are adapted for transferring said plurality of information cells in response to said queue detection means detecting storage of said selected number of said information cells.
15. An interconnect network according to claim 1, wherein said interconnect network is adapted for transferring information as information cells, and said expanded transfer elements further include means for substantially simultaneously transferring a plurality of said information cells out of said expanded interconnect module.
16. An interconnect network according to claim 15 further comprising a memory queue for intermediately storing said plurality of information cells to be transferred, and queue detection means for detecting when a selected number of said information cells are stored in said memory queue, wherein said expanded transfer elements are adapted for transferring said plurality of information cells in response to said queue detection means detecting storage of said selected number of said information cells.
17. An interconnect network according to claim 1, wherein said local I/O channels have associated destination addresses, and said expanded interconnect module includes, array means for storing path information representative of a plurality of paths through said expanded transfer elements over which information from a first local I/O channel of one of said local interconnect modules can be transferred to a second local I/O channel of another one of said local interconnect modules, and index means for selecting appropriate path information from said array means, at least partially in response to a destination address of said second local I/O channel.
18. An interconnect network according to claim 1 , wherein said local transfer elements and said expanded transfer elements are substantially identical.
19. An interconnect network according to claim 16, wherein said local and said expanded transfer elements each include mode control means for selecting whether said transfer element is to be employed in one of said local interconnect modules or in said expanded interconnect module.
20. A dynamically scalable communication interconnect comprising, a selectable number of local interconnects, each having associated transfer elements for transferring information through said associated local interconnect, and a single expanded interconnect, including elements for coupling to said selected number of local interconnects, and expanded transfer elements for transferring information between said local interconnects, wherein said selected number of local interconnects can be varied while said expanded interconnect is operating transferring information.
21. An interconnect network for operation within a communication node, said network comprising, a selectable number of local interconnect modules each having local transfer elements for transferring information between a plurality of local I/O channels and for transferring information between said plurality of local I/O channels and a plurality of non-local I/O channels, wherein said interconnect network is adapted for transferring information as information cells, each of said cells including groups of information words, an expanded interconnect module having coupling means for electrically coupling to said non-local I/O channels, and expanded transfer elements for transferring information between said local interconnect modules, and redundancy generating means for generating an alternative version of information being transferred out of said interconnect network by generating a bit-by-bit "exclusive-or" between pairs of said groups of information words included in an information cell.
22. An interconnect network for operation within a communication node, said network comprising, a selectable number of local interconnect modules each having local transfer elements for transferring information between a plurality of local I/O channels and for transferring information between said plurality of local I/O channels and a plurality of non-local I/O channels, an expanded interconnect module having coupling means for electrically coupling to said non-local I/O channels, and expanded transfer elements for transferring information between said local interconnect modules, and quality of service means for differentiating between information coupled into said local I/O channels based on an associated priority of said information, and for indicating unavailability for receiving information having a particular associated priority on one or more of said local I/O channels.
23. An interconnect network for operation within a communication node, said network comprising, a selectable number of local interconnect modules each having local transfer elements for transferring information between a plurality of local I/O channels and for transferring information between said plurality of local I/O channels and a plurality of non-local I/O channels, and an expanded interconnect module having coupling means for electrically coupling to said non-local I/O channels, and expanded transfer elements for transferring information between said local interconnect modules, wherein said interconnect network is adapted for transferring information as information cells, and said local and expanded transfer elements further include clumping means for substantially simultaneously transferring a plurality of said information cells.
24. A dynamically bandwidth scalable communication node comprising, a selectable number of local communication modules, each of said modules having a plurality of external communication channels for coupling information into and out of said node, a plurality of internal communication channels for coupling information within said node, means for coupling information between said external communication channels and said internal communication channels, a local interconnect module, associated with each of said selected number of local communication modules, and having local transfer elements for transferring information between said plurality of internal communication channels of said associated local communication module, and an expanded interconnect module, including means for coupling to each of said local interconnect modules, and expanded transfer elements for transferring information between said local interconnect modules, thereby enabling transfer of information between external communication channels of different ones of said selectable number of local communication modules.
25. A dynamically bandwidth scalable communication node having a modular construction and comprising, a selectable number of local communication modules, each module having a plurality of communication channels for transferring information into and out of said node, a local interconnect module coupled to each of said local communication modules and including local transfer elements for transferring information between said plurality of communication channels, an expanded interconnect module having elements for coupling to a plurality of said local interconnect modules, and transfer elements for transferring information between said plurality of local interconnect modules, wherein information can be transferred between communication channels of different ones of said local communication modules.
26. A method for transferring information signals between I/O channels of an interconnect network adapted for operation in a communication node, said method comprising the steps of, coupling information to I/O channels of one or more proximately located local interconnect modules, transferring information between I/O channels of a particular local interconnect module in response to a local destination address, transferring information from one of said local interconnect modules to an I/O channel of a proximately located expanded interconnect module in response to a nonlocal destination address, transferring information from said expanded interconnect module to a particular one of said local interconnect modules in response to said non-local destination address, scaling bandwidth of said interconnect network by including a selected number of said local interconnect modules in said plurality of local interconnect modules.
PCT/US1999/013641 1998-06-19 1999-06-18 An interconnect network for operation within a communication node WO1999066758A2 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
IL13461199A IL134611A (en) 1998-06-19 1999-06-18 Interconnect network for operation within a communication node
EP99930340A EP1066735B1 (en) 1998-06-19 1999-06-18 An interconnect network for operation within a communication node
CA002301910A CA2301910A1 (en) 1998-06-19 1999-06-18 An interconnect network for operation within a communication node
AU46896/99A AU760640B2 (en) 1998-06-19 1999-06-18 An interconnect network for operation within a communication node

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US9002898P 1998-06-19 1998-06-19
US60/090,028 1998-06-19

Publications (2)

Publication Number Publication Date
WO1999066758A2 true WO1999066758A2 (en) 1999-12-23
WO1999066758A3 WO1999066758A3 (en) 2000-11-09

Family

ID=22220878

Family Applications (4)

Application Number Title Priority Date Filing Date
PCT/US1999/013857 WO1999066761A1 (en) 1998-06-19 1999-06-18 Device for performing ip forwarding and atm switching
PCT/US1999/013859 WO1999066762A1 (en) 1998-06-19 1999-06-18 Method and system for encapsulating/decapsulating data on a per channel basis in hardware
PCT/US1999/013805 WO1999066675A1 (en) 1998-06-19 1999-06-18 A quality of service facility in a device for performing ip forwarding and atm switching
PCT/US1999/013641 WO1999066758A2 (en) 1998-06-19 1999-06-18 An interconnect network for operation within a communication node

Family Applications Before (3)

Application Number Title Priority Date Filing Date
PCT/US1999/013857 WO1999066761A1 (en) 1998-06-19 1999-06-18 Device for performing ip forwarding and atm switching
PCT/US1999/013859 WO1999066762A1 (en) 1998-06-19 1999-06-18 Method and system for encapsulating/decapsulating data on a per channel basis in hardware
PCT/US1999/013805 WO1999066675A1 (en) 1998-06-19 1999-06-18 A quality of service facility in a device for performing ip forwarding and atm switching

Country Status (8)

Country Link
US (8) US6909720B1 (en)
EP (4) EP1005779B1 (en)
CN (4) CN1214689C (en)
AU (4) AU760640B2 (en)
CA (4) CA2301823C (en)
DE (1) DE69938329T2 (en)
IL (4) IL134612A (en)
WO (4) WO1999066761A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1239694A2 (en) * 2001-03-06 2002-09-11 Redfern Broadband Networks Pty Ltd Optical traffic grooming

Families Citing this family (306)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4481517B2 (en) * 2001-03-19 2010-06-16 株式会社日立製作所 Internetwork apparatus and internetwork method
US6272108B1 (en) * 1997-03-05 2001-08-07 Paradyne Corporation Apparatus and method to allow a frame check sequence to determine the updating of adaptive receiver parameters of a high speed communication device
US6975631B1 (en) * 1998-06-19 2005-12-13 Juniper Networks, Inc. Network packet forwarding lookup with a reduced number of memory accesses
US6980543B1 (en) * 1998-06-19 2005-12-27 Juniper Networks, Inc. Interconnect network for operation within a communication node
EP1005779B1 (en) * 1998-06-19 2008-03-12 Juniper Networks, Inc. Device for performing ip forwarding and atm switching
JP3111993B2 (en) * 1998-07-14 2000-11-27 日本電気株式会社 Node device
US7382736B2 (en) 1999-01-12 2008-06-03 Mcdata Corporation Method for scoring queued frames for selective transmission through a switch
US6952401B1 (en) * 1999-03-17 2005-10-04 Broadcom Corporation Method for load balancing in a network switch
WO2000076152A1 (en) * 1999-06-03 2000-12-14 Fujitsu Network Communications, Inc. Method and system for transmitting traffic in a virtual tunnel of a transmission line
WO2001015386A2 (en) * 1999-08-24 2001-03-01 Telefonaktiebolaget Lm Ericsson (Publ) Differentiated services provisioning for legacy systems
US7046665B1 (en) * 1999-10-26 2006-05-16 Extreme Networks, Inc. Provisional IP-aware virtual paths over networks
US6778536B1 (en) * 1999-11-09 2004-08-17 Synchrodyne Networks, Inc. Combined wavelength division multiplexing, time division multiplexing, and asynchronous packet switching with common time reference
FI110397B (en) * 2000-01-07 2003-01-15 Nokia Corp Procedure for configuring a base station network
US7159235B2 (en) * 2000-01-28 2007-01-02 Sedna Patent Services, Llc Method and apparatus for content distribution via non-homogeneous access networks
US6853618B1 (en) * 2000-02-07 2005-02-08 Marconi Communications, Inc. UPC fail down
JP4741090B2 (en) * 2000-02-09 2011-08-03 ノーテル・ネットワークス・リミテッド Data stream transmission method
US6944163B2 (en) * 2000-02-09 2005-09-13 Nortel Networks Limited 10 Gigabit ethernet mappings for a common LAN/WAN PMD interface with a simple universal physical medium dependent interface
US7126950B2 (en) * 2000-02-14 2006-10-24 Nec Corporation Method and system for transmission and reception of asynchronously multiplexed signals
US6751214B1 (en) * 2000-03-30 2004-06-15 Azanda Network Devices, Inc. Methods and apparatus for dynamically allocating bandwidth between ATM cells and packets
US6785734B1 (en) * 2000-04-10 2004-08-31 International Business Machines Corporation System and method for processing control information from a general through a data processor when a control processor of a network processor being congested
US6862265B1 (en) * 2000-04-13 2005-03-01 Advanced Micro Devices, Inc. Weighted fair queuing approximation in a network switch using weighted round robin and token bucket filter
FR2808150B1 (en) * 2000-04-19 2002-11-29 France Telecom METHOD FOR TESTING A SWITCHING SYSTEM, AND INSERTION DEVICE FOR USE IN THIS METHOD
US6934280B1 (en) * 2000-05-04 2005-08-23 Nokia, Inc. Multiple services emulation over a single network service
US6775229B1 (en) 2000-05-05 2004-08-10 Fujitsu Network Communications, Inc. Method and system for providing a protection path for connection-oriented signals in a telecommunications network
US6693909B1 (en) * 2000-05-05 2004-02-17 Fujitsu Network Communications, Inc. Method and system for transporting traffic in a packet-switched network
US7058730B2 (en) 2000-05-05 2006-06-06 Fujitsu Limited Unique address space and method for a transport network
US7075927B2 (en) 2000-05-05 2006-07-11 Fujitsu Limited Method and system for quality of service (QoS) support in a packet-switched network
US7047176B2 (en) 2000-05-05 2006-05-16 Fujitsu Limited Method and system for hardware simulation
US7133403B1 (en) 2000-05-05 2006-11-07 Fujitsu Limited Transport network and method
US7173912B2 (en) 2000-05-05 2007-02-06 Fujitsu Limited Method and system for modeling and advertising asymmetric topology of a node in a transport network
US6515966B1 (en) 2000-05-05 2003-02-04 Fujitsu Network Communications, Inc. System and method for application object transport
US7151773B1 (en) 2000-05-05 2006-12-19 Fujitsu Limited System and method for connectionless/connection oriented signal transport
US7385917B1 (en) 2000-05-05 2008-06-10 Fujitsu Limited Method and system for providing a protection path for connectionless signals in a telecommunications network
US6801500B1 (en) * 2000-05-18 2004-10-05 Cisco Technology, Inc. Method and apparatus for providing reserved rates to multiple flows on a network interface
US7062642B1 (en) * 2000-05-20 2006-06-13 Ciena Corporation Policy based provisioning of network device resources
JP2001333091A (en) * 2000-05-23 2001-11-30 Fujitsu Ltd Communication equipment
US6785236B1 (en) * 2000-05-28 2004-08-31 Lucent Technologies Inc. Packet transmission scheduling with threshold based backpressure mechanism
GB0016185D0 (en) * 2000-06-30 2000-08-23 Nokia Networks Oy Dynamic DSCP availability request method
JP3687501B2 (en) * 2000-07-05 2005-08-24 日本電気株式会社 Transmission queue management system and management method for packet switch
US7111163B1 (en) 2000-07-10 2006-09-19 Alterwan, Inc. Wide area network using internet with quality of service
US7184440B1 (en) * 2000-07-26 2007-02-27 Alcatel Canada Inc. Multi-protocol switch and method therefore
US7139270B1 (en) * 2000-08-22 2006-11-21 Lucent Technologies Inc. Systems and method for transporting multiple protocol formats in a lightwave communication network
US8619793B2 (en) * 2000-08-21 2013-12-31 Rockstar Consortium Us Lp Dynamic assignment of traffic classes to a priority queue in a packet forwarding device
US20020143987A1 (en) * 2000-08-22 2002-10-03 Sadler Andrew Paul Message management systems and method
US7103063B2 (en) * 2000-08-24 2006-09-05 Tellabs Reston, Inc. Apparatus and method for facilitating data packet transportation
US20020129086A1 (en) * 2000-08-31 2002-09-12 The Regents Of The University Of California Cluster-based aggregated switching technique (CAST) for routing data packets and information objects in computer networks
US20020167967A1 (en) * 2000-09-06 2002-11-14 Schneider Electric Method for managing bandwidth on an ethernet network
US8032653B1 (en) * 2000-09-08 2011-10-04 Juniper Networks, Inc. Guaranteed bandwidth sharing in a traffic shaping system
US7272643B1 (en) 2000-09-13 2007-09-18 Fortinet, Inc. System and method for managing and provisioning virtual routers
US7487232B1 (en) 2000-09-13 2009-02-03 Fortinet, Inc. Switch management system and method
US7574495B1 (en) * 2000-09-13 2009-08-11 Fortinet, Inc. System and method for managing interworking communications protocols
US7389358B1 (en) * 2000-09-13 2008-06-17 Fortinet, Inc. Distributed virtual system to support managed, network-based services
US7444398B1 (en) * 2000-09-13 2008-10-28 Fortinet, Inc. System and method for delivering security services
US8250357B2 (en) 2000-09-13 2012-08-21 Fortinet, Inc. Tunnel interface for securing traffic over a network
US6738392B1 (en) * 2000-09-27 2004-05-18 Cisco Technology, Inc. Method and apparatus of framing high-speed signals
EP1197695A3 (en) * 2000-10-13 2003-04-16 Honda Giken Kogyo Kabushiki Kaisha Spool valve
US7606898B1 (en) 2000-10-24 2009-10-20 Microsoft Corporation System and method for distributed management of shared computers
US6915338B1 (en) * 2000-10-24 2005-07-05 Microsoft Corporation System and method providing automatic policy enforcement in a multi-computer service application
US7113900B1 (en) * 2000-10-24 2006-09-26 Microsoft Corporation System and method for logical modeling of distributed computer systems
CA2426831A1 (en) 2000-10-26 2002-08-01 Wave7 Optics, Inc. Method and system for processing downstream packets of an optical network
US7088737B1 (en) * 2000-10-27 2006-08-08 Redback Networks Inc. Method and apparatus for combining packets having different protocol encapsulations within a circuit
US7133399B1 (en) * 2000-10-31 2006-11-07 Chiaro Networks Ltd System and method for router central arbitration
US7324500B1 (en) * 2000-10-31 2008-01-29 Jeremy Benjamin As Receiver For Chiaro Networks Ltd. Router network protection using multiple facility interfaces
US6993043B1 (en) * 2000-11-13 2006-01-31 At&T Corp. End-to-end prioritized data delivery on networks using IP over frame relay
US7596139B2 (en) 2000-11-17 2009-09-29 Foundry Networks, Inc. Backplane interface adapter with error control and redundant fabric
US7236490B2 (en) 2000-11-17 2007-06-26 Foundry Networks, Inc. Backplane interface adapter
CA2326851A1 (en) * 2000-11-24 2002-05-24 Redback Networks Systems Canada Inc. Policy change characterization method and apparatus
US6987732B2 (en) * 2000-12-15 2006-01-17 Tellabs San Jose, Inc. Apparatus and methods for scheduling packets in a broadband data stream
KR100358153B1 (en) * 2000-12-18 2002-10-25 한국전자통신연구원 QoS supported IP packet forwarding dispersion processing apparatus and method
US7002980B1 (en) 2000-12-19 2006-02-21 Chiaro Networks, Ltd. System and method for router queue and congestion management
US6914883B2 (en) * 2000-12-28 2005-07-05 Alcatel QoS monitoring system and method for a high-speed DiffServ-capable network element
US20020085591A1 (en) * 2001-01-03 2002-07-04 Michael Mesh Fiber optic communication system
US20020085563A1 (en) * 2001-01-03 2002-07-04 Michael Mesh Packet processing method and engine
SE518720C2 (en) * 2001-02-07 2002-11-12 Ericsson Telefon Ab L M Device and method related to traffic control
US7342942B1 (en) * 2001-02-07 2008-03-11 Cortina Systems, Inc. Multi-service segmentation and reassembly device that maintains only one reassembly context per active output port
EP1303089A4 (en) * 2001-02-14 2009-09-09 Panasonic Corp Digital content distribution system and its apparatus
US7006531B2 (en) * 2001-02-21 2006-02-28 Integrated Device Technology, Inc. Method and apparatus for transmitting streamed ingressing data through a switch fabric that provides read requests at an ingress independent request rate
US7230917B1 (en) * 2001-02-22 2007-06-12 Cisco Technology, Inc. Apparatus and technique for conveying per-channel flow control information to a forwarding engine of an intermediate network node
US7286532B1 (en) * 2001-02-22 2007-10-23 Cisco Technology, Inc. High performance interface logic architecture of an intermediate network node
US7110359B1 (en) * 2001-03-05 2006-09-19 Advanced Micro Devices, Inc. System and method for dynamically updating weights of weighted round robin in output queues
JP4187940B2 (en) * 2001-03-06 2008-11-26 株式会社エヌ・ティ・ティ・ドコモ Packet transmission method and system, packet transmission device, reception device, and transmission / reception device
US7342927B1 (en) * 2001-03-09 2008-03-11 Brooktree Broadband Holding, Inc. Systems and methods for transferring various data types across an ATM network
JP3873639B2 (en) * 2001-03-12 2007-01-24 株式会社日立製作所 Network connection device
JP4481518B2 (en) * 2001-03-19 2010-06-16 株式会社日立製作所 Information relay apparatus and transfer method
JP2002281080A (en) * 2001-03-19 2002-09-27 Fujitsu Ltd Packet switch device and multicast transmitting method
US7289522B2 (en) * 2001-03-20 2007-10-30 Verizon Business Global Llc Shared dedicated access line (DAL) gateway routing discrimination
US20030115480A1 (en) * 2001-12-17 2003-06-19 Worldcom, Inc. System, method and apparatus that employ virtual private networks to resist IP QoS denial of service attacks
US6778498B2 (en) * 2001-03-20 2004-08-17 Mci, Inc. Virtual private network (VPN)-aware customer premises equipment (CPE) edge router
US7739497B1 (en) * 2001-03-21 2010-06-15 Verizon Corporate Services Group Inc. Method and apparatus for anonymous IP datagram exchange using dynamic network address translation
US7068645B1 (en) * 2001-04-02 2006-06-27 Cisco Technology, Inc. Providing different QOS to layer-3 datagrams when transported on tunnels
US7203170B2 (en) * 2001-05-01 2007-04-10 Integrated Device Technology, Inc. Network switch port with weighted random early discard
US7286566B1 (en) 2001-05-08 2007-10-23 Cortina Systems, Inc. Multi-service segmentation and reassembly device that maintains reduced number of segmentation contexts
US7006532B1 (en) * 2001-05-23 2006-02-28 Cisco Technology, Inc. Dynamically building encapsulation and decapsulation chains at run time
US20040170166A1 (en) * 2001-05-24 2004-09-02 Ron Cohen Compression methods for packetized sonet/sdh payloads
US20050198379A1 (en) 2001-06-13 2005-09-08 Citrix Systems, Inc. Automatically reconnecting a client across reliable and persistent communication sessions
US20020191621A1 (en) * 2001-06-14 2002-12-19 Cypress Semiconductor Corp. Programmable protocol processing engine for network packet devices
US20020194363A1 (en) * 2001-06-14 2002-12-19 Cypress Semiconductor Corp. Programmable protocol processing engine for network packet devices
US7181547B1 (en) 2001-06-28 2007-02-20 Fortinet, Inc. Identifying nodes in a ring network
US7877014B2 (en) 2001-07-05 2011-01-25 Enablence Technologies Inc. Method and system for providing a return path for signals generated by legacy video service terminals in an optical network
FR2828046B1 (en) * 2001-07-27 2003-10-10 Thales Sa TASK MANAGEMENT METHOD FOR A PACKET SWITCHING ROUTER AUTOMATION PART OF A SECURE PACKET SWITCHED TRANSMISSION NETWORK
US7181530B1 (en) * 2001-07-27 2007-02-20 Cisco Technology, Inc. Rogue AP detection
US20030039226A1 (en) 2001-08-24 2003-02-27 Kwak Joseph A. Physical layer automatic repeat request (ARQ)
US7336680B2 (en) * 2001-09-18 2008-02-26 Scientific-Atlanta, Inc. Multi-carrier frequency-division multiplexing (FDM) architecture for high speed digital service
US20030053493A1 (en) * 2001-09-18 2003-03-20 Joseph Graham Mobley Allocation of bit streams for communication over-multi-carrier frequency-division multiplexing (FDM)
US7881202B2 (en) * 2002-09-25 2011-02-01 Broadcom Corporation System and method for dropping lower priority packets that are slated for wireless transmission
US8500556B2 (en) * 2001-09-28 2013-08-06 Bally Gaming, Inc. Two-wire exchange system
US8543681B2 (en) * 2001-10-15 2013-09-24 Volli Polymer Gmbh Llc Network topology discovery systems and methods
US8868715B2 (en) * 2001-10-15 2014-10-21 Volli Polymer Gmbh Llc Report generation and visualization systems and methods and their use in testing frameworks for determining suitability of a network for target applications
US7310345B2 (en) 2001-11-01 2007-12-18 International Business Machines Corporation Empty indicators for weighted fair queues
US7187684B2 (en) * 2001-11-01 2007-03-06 International Business Machines Corporation Weighted fair queue having extended effective range
US7280474B2 (en) * 2001-11-01 2007-10-09 International Business Machines Corporation Weighted fair queue having adjustable scaling factor
US7317683B2 (en) * 2001-11-01 2008-01-08 International Business Machines Corporation Weighted fair queue serving plural output ports
US7103051B2 (en) * 2001-11-01 2006-09-05 International Business Machines Corporation QoS scheduler and method for implementing quality of service with aging time stamps
AUPR918001A0 (en) * 2001-11-30 2001-12-20 Foursticks Pty Ltd Credit based algorithm for traffic shaping
US7240123B2 (en) * 2001-12-10 2007-07-03 Nortel Networks Limited Distributed routing core
US7379420B2 (en) * 2001-12-28 2008-05-27 Network Equipment Technologies, Inc. Method and apparatus for multiple qualities of service to different network connections of a single network path
US20030123393A1 (en) * 2002-01-03 2003-07-03 Feuerstraeter Mark T. Method and apparatus for priority based flow control in an ethernet architecture
US7038910B1 (en) 2002-01-07 2006-05-02 Wave7 Optics, Inc. System and method for removing heat from a subscriber optical interface
US7984157B2 (en) 2002-02-26 2011-07-19 Citrix Systems, Inc. Persistent and reliable session securely traversing network components using an encapsulating protocol
US7661129B2 (en) 2002-02-26 2010-02-09 Citrix Systems, Inc. Secure traversal of network components
US7257124B2 (en) * 2002-03-20 2007-08-14 International Business Machines Corporation Method and apparatus for improving the fairness of new attaches to a weighted fair queue in a quality of service (QoS) scheduler
US7680043B2 (en) * 2002-03-20 2010-03-16 International Business Machines Corporation Network processor having fast flow queue disable process
US7110411B2 (en) * 2002-03-25 2006-09-19 Erlang Technology, Inc. Method and apparatus for WFQ scheduling using a plurality of scheduling queues to provide fairness, high scalability, and low computation complexity
FR2838898B1 (en) * 2002-04-19 2004-06-18 Cit Alcatel CENTRALIZED SWITCHING AND ROUTING NEEDLE DEVICE
US20120155466A1 (en) 2002-05-06 2012-06-21 Ian Edward Davis Method and apparatus for efficiently processing data packets in a computer network
US7187687B1 (en) 2002-05-06 2007-03-06 Foundry Networks, Inc. Pipeline method and system for switching packets
US7468975B1 (en) 2002-05-06 2008-12-23 Foundry Networks, Inc. Flexible method for processing data packets in a network routing system for enhanced efficiency and monitoring capability
US7623786B2 (en) 2002-05-20 2009-11-24 Enablence Usa Fttx Networks, Inc. System and method for communicating optical signals to multiple subscribers having various bandwidth demands connected to the same optical waveguide
EP1365548A1 (en) * 2002-05-21 2003-11-26 Alcatel Method for encapsulating variable length packets, and related data packet encapsulator and decapsulator
US7376125B1 (en) * 2002-06-04 2008-05-20 Fortinet, Inc. Service processing switch
US7161904B2 (en) 2002-06-04 2007-01-09 Fortinet, Inc. System and method for hierarchical metering in a virtual router based network switch
US7116665B2 (en) * 2002-06-04 2006-10-03 Fortinet, Inc. Methods and systems for a distributed provider edge
US7177311B1 (en) * 2002-06-04 2007-02-13 Fortinet, Inc. System and method for routing traffic through a virtual router-based network switch
US7203192B2 (en) 2002-06-04 2007-04-10 Fortinet, Inc. Network packet steering
US7466697B1 (en) * 2002-07-23 2008-12-16 Atrica Israel Ltd Link multiplexing mechanism utilizing path oriented forwarding
JP3779655B2 (en) * 2002-07-30 2006-05-31 日本電信電話株式会社 Edge node, signal control server, processing method in edge node, processing method in signal control server, communication quality control method, and communication quality control apparatus
CN100448220C (en) * 2002-07-30 2008-12-31 日本电信电话株式会社 Communication control method and system thereof, packet transfer and monitoring method and system
JP2004080102A (en) * 2002-08-09 2004-03-11 Fujitsu Ltd Packet processing apparatus
US20040131072A1 (en) 2002-08-13 2004-07-08 Starent Networks Corporation Communicating in voice and data communications systems
US7096383B2 (en) 2002-08-29 2006-08-22 Cosine Communications, Inc. System and method for virtual router failover in a network routing system
US7590145B2 (en) * 2002-09-17 2009-09-15 Scientific-Atlanta, Inc. Multiplexing octets from a data flow over MPEG packets
US7542471B2 (en) 2002-10-30 2009-06-02 Citrix Systems, Inc. Method of determining path maximum transmission unit
US7616638B2 (en) 2003-07-29 2009-11-10 Orbital Data Corporation Wavefront detection and disambiguation of acknowledgments
US8233392B2 (en) 2003-07-29 2012-07-31 Citrix Systems, Inc. Transaction boundary detection for reduction in timeout penalties
US7630305B2 (en) 2003-07-29 2009-12-08 Orbital Data Corporation TCP selective acknowledgements for communicating delivered and missed data packets
US8270423B2 (en) 2003-07-29 2012-09-18 Citrix Systems, Inc. Systems and methods of using packet boundaries for reduction in timeout prevention
US7266120B2 (en) 2002-11-18 2007-09-04 Fortinet, Inc. System and method for hardware accelerated packet multicast in a virtual routing system
EP1584164A2 (en) * 2002-12-31 2005-10-12 Conexant, Inc. System and method for providing quality of service in asynchronous transfer mode cell transmission
US7983239B1 (en) 2003-01-07 2011-07-19 Raytheon Bbn Technologies Corp. Systems and methods for constructing a virtual model of a multi-hop, multi-access network
US7492714B1 (en) * 2003-02-04 2009-02-17 Pmc-Sierra, Inc. Method and apparatus for packet grooming and aggregation
US8122106B2 (en) 2003-03-06 2012-02-21 Microsoft Corporation Integrating design, deployment, and management phases for systems
US7689676B2 (en) 2003-03-06 2010-03-30 Microsoft Corporation Model-based policy application
US7890543B2 (en) 2003-03-06 2011-02-15 Microsoft Corporation Architecture for distributed computing system and automated design, deployment, and management of distributed applications
US7454141B2 (en) 2003-03-14 2008-11-18 Enablence Usa Fttx Networks Inc. Method and system for providing a return path for signals generated by legacy terminals in an optical network
US7673070B1 (en) * 2003-03-17 2010-03-02 Network Equipment Technologies, Inc. Method of sharing telecommunications node equipment facilities
US20040184071A1 (en) * 2003-03-19 2004-09-23 Garg Man M. Pluggable pre-parser for PJL and PS for extracting print job relates header information from a running stream
AU2004227600B2 (en) * 2003-04-09 2009-05-07 Cisco Technology, Inc. Selective diversion and injection of communication traffic
US6901072B1 (en) 2003-05-15 2005-05-31 Foundry Networks, Inc. System and method for high speed packet transmission implementing dual transmit and receive pipelines
US8005055B2 (en) * 2003-07-23 2011-08-23 Interdigital Technology Corporation Method and apparatus for determining and managing congestion in a wireless communications system
US8213428B2 (en) * 2003-07-24 2012-07-03 International Business Machines Corporation Methods and apparatus for indexing memory of a network processor
US8238241B2 (en) 2003-07-29 2012-08-07 Citrix Systems, Inc. Automatic detection and window virtualization for flow control
US7656799B2 (en) 2003-07-29 2010-02-02 Citrix Systems, Inc. Flow control system architecture
US8432800B2 (en) 2003-07-29 2013-04-30 Citrix Systems, Inc. Systems and methods for stochastic-based quality of service
US7698453B2 (en) 2003-07-29 2010-04-13 Oribital Data Corporation Early generation of acknowledgements for flow control
US8437284B2 (en) 2003-07-29 2013-05-07 Citrix Systems, Inc. Systems and methods for additional retransmissions of dropped packets
WO2005013529A2 (en) * 2003-08-05 2005-02-10 Covaro Networks, Inc. System and method for many-to-many layer 2 aggregation for sonet paths
US7881229B2 (en) * 2003-08-08 2011-02-01 Raytheon Bbn Technologies Corp. Systems and methods for forming an adjacency graph for exchanging network routing data
US7970009B1 (en) * 2003-08-21 2011-06-28 Brixham Solutions Ltd. Method for performing protocol translation in a network switch
US7606927B2 (en) * 2003-08-27 2009-10-20 Bbn Technologies Corp Systems and methods for forwarding data units in a communications network
US7720095B2 (en) 2003-08-27 2010-05-18 Fortinet, Inc. Heterogeneous media packet bridging
US8166204B2 (en) * 2003-08-29 2012-04-24 Raytheon Bbn Technologies Corp. Systems and methods for automatically placing nodes in an ad hoc network
US7852856B2 (en) * 2003-08-29 2010-12-14 Broadcom Corp. System and method for providing pooling or dynamic allocation of connection context data
US7978716B2 (en) 2003-11-24 2011-07-12 Citrix Systems, Inc. Systems and methods for providing a VPN solution
US7496097B2 (en) * 2003-11-11 2009-02-24 Citrix Gateways, Inc. System, apparatus and method for establishing a secured communications link to form a virtual private network at a network protocol layer other than at which packets are filtered
US7522532B2 (en) * 2003-11-24 2009-04-21 At&T Intellectual Property I, L.P. Layer 2/layer 3 interworking via physical loopback
US7417995B2 (en) * 2004-01-20 2008-08-26 Nortel Networks Limited Method and system for frame relay and ethernet service interworking
US7333508B2 (en) * 2004-01-20 2008-02-19 Nortel Networks Limited Method and system for Ethernet and frame relay network interworking
US7406088B2 (en) 2004-01-20 2008-07-29 Nortel Networks Limited Method and system for ethernet and ATM service interworking
US7505466B2 (en) 2004-01-20 2009-03-17 Nortel Networks Limited Method and system for ethernet and ATM network interworking
US8804728B2 (en) * 2004-01-20 2014-08-12 Rockstar Consortium Us Lp Ethernet differentiated services conditioning
US20060196395A1 (en) * 2004-01-21 2006-09-07 Wen-Sheng Lin Foldable table
US7707948B2 (en) * 2004-01-21 2010-05-04 Maxchief Investments, Ltd. Foldable table
US7752982B2 (en) * 2004-01-21 2010-07-13 Maxchief Investments, Ltd. Latching mechanism for foldable table
US7778422B2 (en) 2004-02-27 2010-08-17 Microsoft Corporation Security associations for devices
US7817659B2 (en) 2004-03-26 2010-10-19 Foundry Networks, Llc Method and apparatus for aggregating input data streams
US20050220090A1 (en) * 2004-03-31 2005-10-06 Kevin Loughran Routing architecture
EP1738545A4 (en) * 2004-04-20 2012-04-04 Nortel Networks Ltd Method and system for quality of service support for ethernet multiservice interworking over multiprotocol label switching
US8730961B1 (en) 2004-04-26 2014-05-20 Foundry Networks, Llc System and method for optimizing router lookup
US20050246529A1 (en) 2004-04-30 2005-11-03 Microsoft Corporation Isolated persistent identity storage for authentication of computing devies
US8331377B2 (en) 2004-05-05 2012-12-11 Qualcomm Incorporated Distributed forward link schedulers for multi-carrier communication systems
RU2354061C2 (en) 2004-05-05 2009-04-27 Квэлкомм Инкорпорейтед Method and device for time-delay adaptive control in wireless communication system
US7684404B1 (en) * 2004-06-10 2010-03-23 Intel Corporation Asynchronous transfer mode (ATM) cell formatting
US7757074B2 (en) 2004-06-30 2010-07-13 Citrix Application Networking, Llc System and method for establishing a virtual private network
US8739274B2 (en) 2004-06-30 2014-05-27 Citrix Systems, Inc. Method and device for performing integrated caching in a data communication network
US8495305B2 (en) 2004-06-30 2013-07-23 Citrix Systems, Inc. Method and device for performing caching of dynamically generated objects in a data communication network
CN1305279C (en) * 2004-07-09 2007-03-14 清华大学 Non-state end-to-end constraint entrance permit control method for kernel network
EP1771979B1 (en) 2004-07-23 2011-11-23 Citrix Systems, Inc. A method and systems for securing remote access to private networks
EP1771998B1 (en) 2004-07-23 2015-04-15 Citrix Systems, Inc. Systems and methods for optimizing communications between network nodes
CN1294728C (en) * 2004-08-05 2007-01-10 华为技术有限公司 Method and system for providing QoS assurance in edge router
CA2576944A1 (en) * 2004-08-10 2006-02-23 Wave7 Optics, Inc. Countermeasures for idle pattern srs interference in ethernet optical network systems
US7657657B2 (en) 2004-08-13 2010-02-02 Citrix Systems, Inc. Method for maintaining transaction integrity across multiple remote access servers
US7102194B2 (en) * 2004-08-16 2006-09-05 System General Corp. High voltage and low on-resistance LDMOS transistor having radiation structure and isolation effect
US7599622B2 (en) 2004-08-19 2009-10-06 Enablence Usa Fttx Networks Inc. System and method for communicating optical signals between a data service provider and subscribers
US7499419B2 (en) * 2004-09-24 2009-03-03 Fortinet, Inc. Scalable IP-services enabled multicast forwarding with efficient resource utilization
US7606158B2 (en) * 2004-09-24 2009-10-20 Cisco Technology, Inc. Hierarchical flow control for router ATM interfaces
US7748032B2 (en) 2004-09-30 2010-06-29 Citrix Systems, Inc. Method and apparatus for associating tickets in a ticket hierarchy
US8613048B2 (en) 2004-09-30 2013-12-17 Citrix Systems, Inc. Method and apparatus for providing authorized remote access to application sessions
US7711835B2 (en) 2004-09-30 2010-05-04 Citrix Systems, Inc. Method and apparatus for reducing disclosure of proprietary data in a networked environment
CN100466597C (en) * 2004-10-11 2009-03-04 华为技术有限公司 Method for implementing allocation of traffic flow based on different service domains
US7657703B1 (en) 2004-10-29 2010-02-02 Foundry Networks, Inc. Double density content addressable memory (CAM) lookup scheme
US7808904B2 (en) 2004-11-18 2010-10-05 Fortinet, Inc. Method and apparatus for managing subscriber profiles
US7406085B2 (en) * 2004-11-19 2008-07-29 Analog Devices, Inc. Auto configuration for asynchronous transfer mode based access device
US7460490B2 (en) * 2004-11-19 2008-12-02 Analog Devices, Inc. Auto configuration for asynchronous transfer mode based access device
US7434192B2 (en) * 2004-12-13 2008-10-07 Altera Corporation Techniques for optimizing design of a hard intellectual property block for data transmission
US7810089B2 (en) 2004-12-30 2010-10-05 Citrix Systems, Inc. Systems and methods for automatic installation and execution of a client-side acceleration program
US8954595B2 (en) 2004-12-30 2015-02-10 Citrix Systems, Inc. Systems and methods for providing client-side accelerated access to remote applications via TCP buffering
US8706877B2 (en) 2004-12-30 2014-04-22 Citrix Systems, Inc. Systems and methods for providing client-side dynamic redirection to bypass an intermediary
US8549149B2 (en) 2004-12-30 2013-10-01 Citrix Systems, Inc. Systems and methods for providing client-side accelerated access to remote applications via TCP multiplexing
US8700695B2 (en) 2004-12-30 2014-04-15 Citrix Systems, Inc. Systems and methods for providing client-side accelerated access to remote applications via TCP pooling
US7581005B2 (en) * 2005-01-20 2009-08-25 Citrix Systems, Inc. Systems and methods for preserving transport layer protocol options
US8077632B2 (en) * 2005-01-20 2011-12-13 Citrix Systems, Inc. Automatic LAN/WAN port detection
JP5183214B2 (en) 2005-01-24 2013-04-17 サイトリックス システムズ,インコーポレイテッド System and method for performing caching of dynamically generated objects in a network
US8255456B2 (en) 2005-12-30 2012-08-28 Citrix Systems, Inc. System and method for performing flash caching of dynamically generated objects in a data communication network
US8024568B2 (en) 2005-01-28 2011-09-20 Citrix Systems, Inc. Method and system for verification of an endpoint security scan
JP2006279514A (en) * 2005-03-29 2006-10-12 Fujitsu Ltd Device and method for controlling communication
CN100452766C (en) * 2005-04-12 2009-01-14 中国电信股份有限公司 IP data stream guiding system and method based on business trigging
US8489728B2 (en) 2005-04-15 2013-07-16 Microsoft Corporation Model-based system monitoring
US7802144B2 (en) 2005-04-15 2010-09-21 Microsoft Corporation Model-based system monitoring
US7797147B2 (en) 2005-04-15 2010-09-14 Microsoft Corporation Model-based system monitoring
US20060268943A1 (en) * 2005-05-16 2006-11-30 International Business Machines Corporation Methods, systems, and storage media for data encapsulation in networks
CA2611175A1 (en) * 2005-06-06 2006-12-14 Mobidia, Inc. Data packet structure and protocol
US8549513B2 (en) 2005-06-29 2013-10-01 Microsoft Corporation Model-based virtual system provisioning
JP4564433B2 (en) * 2005-09-26 2010-10-20 富士通株式会社 Transmission apparatus and frame transfer method
US20070111698A1 (en) * 2005-10-26 2007-05-17 Manamohan Mysore Method and apparatus for providing bearer selection and transmission parameter configuration
US7941309B2 (en) 2005-11-02 2011-05-10 Microsoft Corporation Modeling IT operations/policies
US8448162B2 (en) 2005-12-28 2013-05-21 Foundry Networks, Llc Hitless software upgrades
US7921184B2 (en) 2005-12-30 2011-04-05 Citrix Systems, Inc. System and method for performing flash crowd caching of dynamically generated objects in a data communication network
US8301839B2 (en) 2005-12-30 2012-10-30 Citrix Systems, Inc. System and method for performing granular invalidation of cached dynamically generated objects in a data communication network
US20070171825A1 (en) * 2006-01-20 2007-07-26 Anagran, Inc. System, method, and computer program product for IP flow routing
US8547843B2 (en) * 2006-01-20 2013-10-01 Saisei Networks Pte Ltd System, method, and computer program product for controlling output port utilization
US8880746B2 (en) * 2006-01-31 2014-11-04 Cisco Technology, Inc. Method using virtual ASIC pins to reconfigure hardware
US7668920B2 (en) * 2006-03-01 2010-02-23 Fortinet, Inc. Electronic message and data tracking system
US20070237074A1 (en) * 2006-04-06 2007-10-11 Curry David S Configuration of congestion thresholds for a network traffic management system
US7881314B2 (en) * 2006-05-01 2011-02-01 Cisco Technology, Inc. Network device providing access to both layer 2 and layer 3 services on a single physical interface
TWI429219B (en) * 2006-05-01 2014-03-01 Koninkl Philips Electronics Nv Method of reserving resources with a maximum delay guarantee for multi-hop transmission in a distributed access wireless communications network
US7756134B2 (en) 2006-05-02 2010-07-13 Harris Corporation Systems and methods for close queuing to support quality of service
US7894509B2 (en) * 2006-05-18 2011-02-22 Harris Corporation Method and system for functional redundancy based quality of service
US7990860B2 (en) * 2006-06-16 2011-08-02 Harris Corporation Method and system for rule-based sequencing for QoS
US7856012B2 (en) * 2006-06-16 2010-12-21 Harris Corporation System and methods for generic data transparent rules to support quality of service
US8516153B2 (en) 2006-06-16 2013-08-20 Harris Corporation Method and system for network-independent QoS
US8064464B2 (en) 2006-06-16 2011-11-22 Harris Corporation Method and system for inbound content-based QoS
US7916626B2 (en) 2006-06-19 2011-03-29 Harris Corporation Method and system for fault-tolerant quality of service
US8730981B2 (en) * 2006-06-20 2014-05-20 Harris Corporation Method and system for compression based quality of service
US7769028B2 (en) 2006-06-21 2010-08-03 Harris Corporation Systems and methods for adaptive throughput management for event-driven message-based data
US8149797B2 (en) * 2006-06-30 2012-04-03 Hewlett-Packard Development Company, L.P. Visualization of RF neighbor correlation in a single view
US20080075003A1 (en) * 2006-09-21 2008-03-27 Futurewei Technologies, Inc. Method and system for admission and congestion control of network communication traffic
JP4701152B2 (en) * 2006-10-20 2011-06-15 富士通株式会社 Data relay apparatus, data relay method, and data relay program
US8533846B2 (en) 2006-11-08 2013-09-10 Citrix Systems, Inc. Method and system for dynamically associating access rights with a resource
US8238255B2 (en) 2006-11-22 2012-08-07 Foundry Networks, Llc Recovering from failures without impact on data traffic in a shared bus architecture
US7725594B2 (en) * 2006-12-29 2010-05-25 Verizon Patent And Licensing Inc. Assigning priority to network traffic at customer premises
US8155011B2 (en) 2007-01-11 2012-04-10 Foundry Networks, Llc Techniques for using dual memory structures for processing failure detection protocol packets
US7664857B2 (en) 2007-01-26 2010-02-16 Citrix Systems, Inc. Systems and methods of using an IP ID field for automatic WAN/LAN detection
US7760642B2 (en) 2007-03-12 2010-07-20 Citrix Systems, Inc. Systems and methods for providing quality of service precedence in TCP congestion control
US8631147B2 (en) * 2007-03-12 2014-01-14 Citrix Systems, Inc. Systems and methods for configuring policy bank invocations
US7796510B2 (en) 2007-03-12 2010-09-14 Citrix Systems, Inc. Systems and methods for providing virtual fair queueing of network traffic
CN101136856B (en) * 2007-06-05 2011-06-22 中兴通讯股份有限公司 Inter-board combined packet transmission method and system
US8037399B2 (en) 2007-07-18 2011-10-11 Foundry Networks, Llc Techniques for segmented CRC design in high speed networks
US8271859B2 (en) 2007-07-18 2012-09-18 Foundry Networks Llc Segmented CRC design in high speed networks
US8149839B1 (en) 2007-09-26 2012-04-03 Foundry Networks, Llc Selection of trunk ports and paths using rotation
US20090109960A1 (en) * 2007-10-30 2009-04-30 Motorola, Inc. Method and apparatus for a virtual circuit data area within a packet data frame
US8780762B2 (en) 2009-03-11 2014-07-15 Sony Corporation Node query in ad hoc home mesh network
US8861445B2 (en) * 2009-03-11 2014-10-14 Sony Cororation Multi-channel single radio communication in home mesh network
US20100232400A1 (en) * 2009-03-11 2010-09-16 Sony Corporation Virtualizing single radio for multiple wireless interfaces in home mesh network
US7974297B2 (en) * 2009-03-11 2011-07-05 Sony Corporation Quality of service queue management in home mesh network
US8223786B2 (en) * 2009-03-11 2012-07-17 Sony Corporation Quality of service scheduling for home mesh network
US8761174B2 (en) * 2009-03-11 2014-06-24 Sony Corporation Quality of service traffic recognition and packet classification home mesh network
US8194593B2 (en) 2009-03-11 2012-06-05 Sony Corporation Quality of service architecture for home mesh network
US9106539B2 (en) 2009-03-26 2015-08-11 At&T Intellectual Property I, L.P. User-controlled network configuration for handling multiple classes of service
KR101460848B1 (en) 2009-04-01 2014-11-20 니시라, 인크. Method and apparatus for implementing and managing virtual switches
US8090901B2 (en) 2009-05-14 2012-01-03 Brocade Communications Systems, Inc. TCAM management approach that minimize movements
CN101998679B (en) * 2009-08-13 2012-11-07 华为技术有限公司 Transmission bearing relaying method, device and communication system
US8599850B2 (en) 2009-09-21 2013-12-03 Brocade Communications Systems, Inc. Provisioning single or multistage networks using ethernet service instances (ESIs)
WO2011047373A1 (en) * 2009-10-18 2011-04-21 Tellabs Operations, Inc. Method and apparatus for increasing overall aggregate capacity of a network
US8537846B2 (en) * 2010-04-27 2013-09-17 Hewlett-Packard Development Company, L.P. Dynamic priority queue level assignment for a network flow
US8537669B2 (en) 2010-04-27 2013-09-17 Hewlett-Packard Development Company, L.P. Priority queue level optimization for a network flow
CN101854717A (en) * 2010-06-09 2010-10-06 中兴通讯股份有限公司 Data transmission method, user equipment and base station
US20130060901A1 (en) * 2010-06-16 2013-03-07 Ervin Adrovic System for information management protection and routing
US8850089B1 (en) * 2010-06-18 2014-09-30 Integrated Device Technology, Inc. Method and apparatus for unified final buffer with pointer-based and page-based scheme for traffic optimization
US8966040B2 (en) * 2010-07-06 2015-02-24 Nicira, Inc. Use of network information base structure to establish communication between applications
US8817594B2 (en) * 2010-07-13 2014-08-26 Telefonaktiebolaget L M Ericsson (Publ) Technique establishing a forwarding path in a network system
CN102377633B (en) * 2010-08-06 2014-10-08 北京乾唐视联网络科技有限公司 Communication connection method and system of access network device
CN102137008B (en) * 2011-01-24 2013-08-07 华为技术有限公司 Quality of service (QoS) keeping method, device and system
US9225656B2 (en) 2011-02-07 2015-12-29 Brocade Communications Systems, Inc. Quality of service in a heterogeneous network
US8787351B2 (en) * 2011-04-14 2014-07-22 Alcatel Lucent Method and apparatus for scheduling transmissions in a communication network
US9379935B2 (en) * 2011-07-29 2016-06-28 Verizon Patent And Licensing Inc. Cached routing service
JP2014003525A (en) * 2012-06-20 2014-01-09 Nec Commun Syst Ltd Network control device, network control method, and network control program
CN102857564B (en) * 2012-08-27 2016-08-10 华为终端有限公司 A kind of method of data synchronization, terminal and system
US10270564B2 (en) * 2013-03-12 2019-04-23 Huawei Technologies Co., Ltd. System and method for multi-layer protocol selection
EP2802111A4 (en) * 2013-03-12 2015-10-07 Huawei Tech Co Ltd Communication method, device and system
CN104750657A (en) * 2013-12-31 2015-07-01 中国石油化工股份有限公司 Numerical simulation redundancy parallel computing method applicable to fracture-cavity type structure carbonate reservoirs
US10834065B1 (en) 2015-03-31 2020-11-10 F5 Networks, Inc. Methods for SSL protected NTLM re-authentication and devices thereof
US20170188267A1 (en) * 2015-12-29 2017-06-29 Devanand Palanisamy Method And Apparatus For Network Bandwidth Measurement
US10404698B1 (en) 2016-01-15 2019-09-03 F5 Networks, Inc. Methods for adaptive organization of web application access points in webtops and devices thereof
US10263807B2 (en) * 2016-12-19 2019-04-16 Ciena Corporation Hierarchical statistics acceleration
US20180278459A1 (en) * 2017-03-27 2018-09-27 Cisco Technology, Inc. Sharding Of Network Resources In A Network Policy Platform
CN111131059A (en) * 2019-12-04 2020-05-08 视联动力信息技术股份有限公司 Data transmission control method, device and computer readable storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5740156A (en) 1986-09-16 1998-04-14 Hitachi, Ltd. Packet switching system having self-routing switches

Family Cites Families (77)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4998242A (en) * 1988-12-09 1991-03-05 Transwitch Corp. Virtual tributary cross connect switch and switch network utilizing the same
US5081654A (en) * 1989-05-12 1992-01-14 Alcatel Na Network Systems Corp. Parallel bit detection circuit for detecting frame synchronization information imbedded within a serial bit stream and method for carrying out same
JP2821255B2 (en) * 1990-09-12 1998-11-05 株式会社日立製作所 Communication network connection device
US5703879A (en) 1991-08-02 1997-12-30 Gpt Limited ATM switching arrangement
US5255264A (en) 1991-09-26 1993-10-19 Ipc Information Systems, Inc. Distributed control switching network for multi-line telephone communications
US5278824A (en) * 1992-06-02 1994-01-11 At&T Bell Laboratories Dual hubbing in a bidirectional line-switched ring transmission system
IT1259036B (en) * 1992-07-24 1996-03-11 Italtel Spa LOCAL AND / OR TRANSIT CENTRAL UNIT FOR A BROADBAND COMMUNICATION NETWORK
US5490252A (en) * 1992-09-30 1996-02-06 Bay Networks Group, Inc. System having central processor for transmitting generic packets to another processor to be altered and transmitting altered packets back to central processor for routing
US5367520A (en) 1992-11-25 1994-11-22 Bell Communcations Research, Inc. Method and system for routing cells in an ATM switch
US5555244A (en) 1994-05-19 1996-09-10 Integrated Network Corporation Scalable multimedia network
CA2124610C (en) * 1994-05-30 1998-10-06 Charles Kevin Huscroft Integrated user network interface device
US5600653A (en) * 1994-09-30 1997-02-04 Comsat Corporation Technique for improving asynchronous transfer mode operation over a communications link with bursty bit errors
US5802105A (en) 1994-11-30 1998-09-01 Qualcomm Incorporated Method and apparatus for testing a digital communication channel
US5533018A (en) 1994-12-21 1996-07-02 National Semiconductor Corporation Multi-protocol packet framing over an isochronous network
JP3515263B2 (en) * 1995-05-18 2004-04-05 株式会社東芝 Router device, data communication network system, node device, data transfer method, and network connection method
US5729546A (en) * 1995-06-21 1998-03-17 Cisco Systems, Inc. Expandable communication cell bus for multiplexing and concentrating communication cell traffic onto high speed lines
US5751709A (en) 1995-12-28 1998-05-12 Lucent Technologies Inc. Adaptive time slot scheduling apparatus and method for end-points in an ATM network
US5892924A (en) * 1996-01-31 1999-04-06 Ipsilon Networks, Inc. Method and apparatus for dynamically shifting between routing and switching packets in a transmission network
CA2243359A1 (en) 1996-01-31 1997-08-07 Ipsilon Networks, Inc. Improved method and apparatus for dynamically shifting between routing and switching packets in a transmission network
FI2408U1 (en) 1996-02-02 1996-04-29 Tecwill Oy concrete mixing plant
US6021263A (en) * 1996-02-16 2000-02-01 Lucent Technologies, Inc. Management of ATM virtual circuits with resources reservation protocol
US6237029B1 (en) * 1996-02-26 2001-05-22 Argosystems, Inc. Method and apparatus for adaptable digital protocol processing
GB9604619D0 (en) 1996-03-04 1996-05-01 Plessey Telecomm Combined multiplexer
DE69736340T2 (en) * 1996-03-19 2007-07-12 Lucent Technologies Inc. Method and device in an integrated telecommunications network for converting synchronous narrowband signals into signals operating in asynchronous transmission mode
US6219728B1 (en) 1996-04-22 2001-04-17 Nortel Networks Limited Method and apparatus for allocating shared memory resources among a plurality of queues each having a threshold value therefor
FI101921B (en) 1996-06-07 1998-09-15 Nokia Telecommunications Oy Implementation of ATM virtual channels in a packet network
US5764645A (en) * 1996-06-12 1998-06-09 Microsoft Corporation IP/ATM network adaptation
US5936965A (en) * 1996-07-08 1999-08-10 Lucent Technologies, Inc. Method and apparatus for transmission of asynchronous, synchronous, and variable length mode protocols multiplexed over a common bytestream
US6122281A (en) * 1996-07-22 2000-09-19 Cabletron Systems, Inc. Method and apparatus for transmitting LAN data over a synchronous wide area network
FI104670B (en) * 1996-09-24 2000-04-14 Nokia Networks Oy Packet routing in a data communication system
US6463477B1 (en) 1996-09-27 2002-10-08 Mci Communications Corporation Detection of presence of multiprotocol encapsulation in a data packet
US6052373A (en) 1996-10-07 2000-04-18 Lau; Peter S. Y. Fault tolerant multicast ATM switch fabric, scalable speed and port expansion configurations
US5828844A (en) * 1996-10-08 1998-10-27 At&T Corp. Internet NCP over ATM
JPH10126419A (en) 1996-10-23 1998-05-15 Nec Corp Atm exchange system
JP3123447B2 (en) 1996-11-13 2001-01-09 日本電気株式会社 Switch control circuit of ATM exchange
US6097722A (en) * 1996-12-13 2000-08-01 Nortel Networks Corporation Bandwidth management processes and systems for asynchronous transfer mode networks using variable virtual paths
US5903559A (en) * 1996-12-20 1999-05-11 Nec Usa, Inc. Method for internet protocol switching over fast ATM cell transport
US6002692A (en) 1996-12-30 1999-12-14 Hyundai Electronics America Line interface unit for adapting broad bandwidth network to lower bandwidth network fabric
US6047002A (en) * 1997-01-16 2000-04-04 Advanced Micro Devices, Inc. Communication traffic circle system and method for performing packet conversion and routing between different packet formats including an instruction field
US6115373A (en) * 1997-01-24 2000-09-05 The Hong Kong University Of Science And Technology Information network architecture
JP3765899B2 (en) * 1997-01-27 2006-04-12 富士通株式会社 Transmission equipment
JPH10233745A (en) * 1997-02-18 1998-09-02 Nec Corp Multiplex transmission method and system
JPH10242989A (en) 1997-02-28 1998-09-11 Nec Corp Photoelectric composite type network node control system
US6205154B1 (en) * 1997-04-15 2001-03-20 Lucent Technologies, Inc. Automatic path selection for fiber-optic transmission networks
US6134238A (en) * 1997-05-06 2000-10-17 Lucent Technologies Inc. Layered bandwidth management in ATM/SDH (SONET) networks
US5940389A (en) 1997-05-12 1999-08-17 Computer And Communication Research Laboratories Enhanced partially self-routing algorithm for controller Benes networks
US6075788A (en) * 1997-06-02 2000-06-13 Lsi Logic Corporation Sonet physical layer device having ATM and PPP interfaces
US6052364A (en) 1997-06-13 2000-04-18 Comsat Corporation CDMA system architecture for satcom terminals
US6408005B1 (en) 1997-09-05 2002-06-18 Nec Usa, Inc. Dynamic rate control scheduler for ATM networks
DE19740107A1 (en) * 1997-09-12 1999-03-18 Alsthom Cge Alcatel Method for transmitting data packets and network element suitable for carrying out the method
US6223301B1 (en) 1997-09-30 2001-04-24 Compaq Computer Corporation Fault tolerant memory
US6389453B1 (en) * 1997-10-09 2002-05-14 Mci Communications Corporation Method and system for routing undirectional multicast data
JP3561126B2 (en) 1997-10-20 2004-09-02 富士通株式会社 Connectionless communication method
US6198751B1 (en) * 1997-11-19 2001-03-06 Cabletron Systems, Inc. Multi-protocol packet translator
US6052375A (en) 1997-11-26 2000-04-18 International Business Machines Corporation High speed internetworking traffic scaler and shaper
JP3152293B2 (en) * 1997-12-17 2001-04-03 日本電気株式会社 IP layer processing device
US6049542A (en) 1997-12-31 2000-04-11 Samsung Electronics Co., Ltd. Scalable multistage interconnection network architecture and method for performing in-service upgrade thereof
US6125112A (en) 1998-03-23 2000-09-26 3Com Corporation Non-buffered, non-blocking multistage ATM switch
CA2295065A1 (en) * 1998-04-29 1999-11-04 Shang-Tse Chuang Packet-switch system
US6667956B2 (en) * 1998-05-01 2003-12-23 Nortel Networks Limited Multi-class network
US7100020B1 (en) * 1998-05-08 2006-08-29 Freescale Semiconductor, Inc. Digital communications processor
US6205150B1 (en) 1998-05-28 2001-03-20 3Com Corporation Method of scheduling higher and lower priority data packets
US6185635B1 (en) * 1998-05-30 2001-02-06 Alcatel Networks Corporation Method and circuit for transporting data based on the content of ingress data words and egress data words
US6266333B1 (en) * 1998-06-02 2001-07-24 Lucent Technologies Inc. Network-independent routing of communication signals
US6463096B1 (en) 1998-06-12 2002-10-08 Cisco Systems, Inc MAC protocol employing multiple data rates
US6195346B1 (en) * 1998-06-16 2001-02-27 Mci Communications Corporation Method and system for processing an HDLC message
US6487198B1 (en) * 1998-06-16 2002-11-26 Mci Communications Corporation Method and system for unloading T1 payloads from ATM cells
EP1005779B1 (en) 1998-06-19 2008-03-12 Juniper Networks, Inc. Device for performing ip forwarding and atm switching
US6980543B1 (en) 1998-06-19 2005-12-27 Juniper Networks, Inc. Interconnect network for operation within a communication node
US6975631B1 (en) 1998-06-19 2005-12-13 Juniper Networks, Inc. Network packet forwarding lookup with a reduced number of memory accesses
US6466976B1 (en) * 1998-12-03 2002-10-15 Nortel Networks Limited System and method for providing desired service policies to subscribers accessing the internet
US6477168B1 (en) * 1999-02-03 2002-11-05 International Business Machines Corporation Cell/frame scheduling method and communications cell/frame scheduler
US6331978B1 (en) * 1999-03-09 2001-12-18 Nokia Telecommunications, Oy Generic label encapsulation protocol for carrying label switched packets over serial links
US7065037B1 (en) * 1999-03-25 2006-06-20 Samsung Electronics Co., Ltd. Method and apparatus to provide facility and module redundancy in telecommunication switching equipment
US6498792B1 (en) * 1999-06-03 2002-12-24 Fujitsu Network Communications, Inc. Method and apparatus for switching signals of multiple different communication protocols
US6771663B1 (en) * 2000-02-23 2004-08-03 Cypress Semiconductor Corp. Hybrid data transport scheme over optical networks
US6466591B1 (en) * 2000-12-30 2002-10-15 Redback Networks Inc. Method and apparatus for processing of multiple protocols within data and control channels in data transmission signals

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5740156A (en) 1986-09-16 1998-04-14 Hitachi, Ltd. Packet switching system having self-routing switches

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1239694A2 (en) * 2001-03-06 2002-09-11 Redfern Broadband Networks Pty Ltd Optical traffic grooming
EP1239694A3 (en) * 2001-03-06 2002-09-25 Redfern Broadband Networks Pty Ltd Optical traffic grooming

Also Published As

Publication number Publication date
EP1005780A1 (en) 2000-06-07
CA2301853A1 (en) 1999-12-23
CN1275299A (en) 2000-11-29
CN1275283A (en) 2000-11-29
US20100322242A1 (en) 2010-12-23
US20050201387A1 (en) 2005-09-15
US8018947B2 (en) 2011-09-13
US20100020802A1 (en) 2010-01-28
AU760840B2 (en) 2003-05-22
US6611522B1 (en) 2003-08-26
IL134612A (en) 2004-06-01
CN1275298A (en) 2000-11-29
EP1005742A1 (en) 2000-06-07
US6658021B1 (en) 2003-12-02
DE69938329D1 (en) 2008-04-24
AU4689699A (en) 2000-01-05
CN100385876C (en) 2008-04-30
WO1999066675A1 (en) 1999-12-23
EP1066735B1 (en) 2011-08-17
AU4697599A (en) 2000-01-05
DE69938329T2 (en) 2009-03-26
EP1066735A2 (en) 2001-01-10
CN1166247C (en) 2004-09-08
CA2301823A1 (en) 1999-12-23
CA2301736A1 (en) 1999-12-23
US7586919B2 (en) 2009-09-08
US20130238810A1 (en) 2013-09-12
CA2301823C (en) 2006-11-14
CA2301853C (en) 2007-01-09
AU4697399A (en) 2000-01-05
US8432921B2 (en) 2013-04-30
CN1214689C (en) 2005-08-10
WO1999066762A1 (en) 1999-12-23
IL134615A (en) 2004-03-28
AU760313B2 (en) 2003-05-15
CA2301736C (en) 2007-03-06
US7809015B1 (en) 2010-10-05
IL134611A0 (en) 2001-04-30
EP1005779B1 (en) 2008-03-12
EP1005779A1 (en) 2000-06-07
EP1005780B1 (en) 2012-05-02
US6909720B1 (en) 2005-06-21
IL134615A0 (en) 2001-04-30
IL134611A (en) 2004-05-12
CN1284409C (en) 2006-11-08
AU4956499A (en) 2000-01-05
IL134616A (en) 2004-05-12
WO1999066758A3 (en) 2000-11-09
IL134612A0 (en) 2001-04-30
CN1286886A (en) 2001-03-07
US9077777B2 (en) 2015-07-07
CA2301910A1 (en) 1999-12-23
AU760640B2 (en) 2003-05-22
WO1999066761A1 (en) 1999-12-23
AU771091B2 (en) 2004-03-11
IL134616A0 (en) 2001-04-30

Similar Documents

Publication Publication Date Title
AU760640B2 (en) An interconnect network for operation within a communication node
US8306028B2 (en) Interconnect network for operation within a communication node
US6621828B1 (en) Fused switch core and method for a telecommunications node
US7139271B1 (en) Using an embedded indication of egress application type to determine which type of egress processing to perform
US6822960B1 (en) Asynchronous transfer mode (ATM) switch and method
US7286566B1 (en) Multi-service segmentation and reassembly device that maintains reduced number of segmentation contexts
US6631130B1 (en) Method and apparatus for switching ATM, TDM, and packet data through a single communications switch while maintaining TDM timing
US20030152182A1 (en) Optical exchange method, apparatus and system for facilitating data transport between WAN, SAN and LAN and for enabling enterprise computing into networks
US6760327B1 (en) Rate adjustable backplane and method for a telecommunications node
US6628657B1 (en) Method and system for transporting synchronous and asynchronous traffic on a bus of a telecommunications node
US6920156B1 (en) Method and system for transporting synchronous and asynchronous traffic on a synchronous bus of a telecommunications node
US6778529B1 (en) Synchronous switch and method for a telecommunications node
EP2045982A2 (en) Method and apparatus for switching ATM, TDM and packet data through a single communication switch
US6636511B1 (en) Method of multicasting data through a communications switch

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 134611

Country of ref document: IL

Ref document number: 99801400.1

Country of ref document: CN

AK Designated states

Kind code of ref document: A2

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW SD SL SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
ENP Entry into the national phase

Ref document number: 2301910

Country of ref document: CA

Ref document number: 2301910

Country of ref document: CA

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 46896/99

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 1999930340

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

AK Designated states

Kind code of ref document: A3

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A3

Designated state(s): GH GM KE LS MW SD SL SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

WWP Wipo information: published in national office

Ref document number: 1999930340

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 46896/99

Country of ref document: AU