Patent application number | Description | Published |
20080205377 | System and methods for providing server virtualization assistance - An improved system and method for network switching that provides the ability to automatically and seamlessly migrate policies for network hosts under certain conditions to local or remote switch ports. The improved switch automatically migrates switch policies for virtual and physical hosts from a source port to a destination port when hosts have been physically moved or replaced. For elements that have failed, the improved switch also migrates host network policies for both physical hosts and virtual machines when the old host network policy will maintain serviceability of the element when applied to the new port. | 08-28-2008 |
20080215910 | High-Availability Networking with Intelligent Failover - Methods and systems for maintaining high-availability in a computer network using intelligent failover are presented. In a network switch running an OSI model layer-2 or higher protocol on its external links, the protocol state information is monitored to determine failover status of the link to avoid identifying external link failures due to link flapping. One such protocol is the spanning tree protocol. Additionally, flexibility in failover is provided using configurable triggers to define external failure events. The triggers initiate a link drop of one or more internal links of the network switch in response to an external failure event. The link drops, in turn, initiate failover of an attached computing device to a redundant link through a network interface teaming/failover arrangement whereby the computing device switches to an alternative network interface accessing the network through a redundant path. Failover can be selective depending upon VLAN and trunking configurations. | 09-04-2008 |
20080275975 | Blade Server System with at Least One Rack-Switch Having Multiple Switches Interconnected and Configured for Management and Operation as a Single Virtual Switch - Described is a rack-switch including a rack and a plurality of blade server chassis within the rack. Each blade server chassis has a plurality of server blades in communication with at least one switch. Each switch includes a plurality of external ports. At least two of the external ports of each switch are inter-switch link (ISL) ports. The rack-switch also includes a plurality of inter-switch links. Each inter-switch link electrically connects one ISL port of one of the switches to one ISL port of another of the switches. The plurality of inter-switch links interconnects the switches such that the switches are daisy chained in a loop. The inter-switch links convey Ethernet packets representing server-to-server communications between server blades of different blade server chassis. | 11-06-2008 |
20100054260 | Method and Apparatus to Switch Packets between Virtual Ports - A method and network switch for switching data units assigns a unique virtual port to each end-node operating on a physical machine connected to a physical port of network switch. A data unit, sent by a given end-node operating on the physical machine, is received at the physical port. The received data unit is switched to the virtual port assigned to the given end-node. Based on the virtual port assigned to the given end-node, the data unit is switched to a second physical port of the network switch for subsequent forwarding of the data unit towards its destination. | 03-04-2010 |
20100232443 | Method and Apparatus for Managing, Configuring, and Controlling an I/O Virtualization Device through a Network Switch - An apparatus and method bridges frames between entities running on the same server. The server has a physical network interface that supports I/O virtualization. The physical network interface has an IOV device with a bridging function capable of bridging frames between entities running on the server. A network switch has a physical port coupled to the server by a physical link and a switching fabric in communication with the physical port for transmitting and receiving Ethernet data frames to and from the IOV device. A control processor, in communication with the IOV device, configures the bridging function of the IOV device to bridge some frames locally within the server between the entities running on the server and to pass other frames through the IOV device over the physical link to the physical port and the switching fabric, so that the switching fabric bridges these other frames between the entities. | 09-16-2010 |
20110035494 | NETWORK VIRTUALIZATION FOR A VIRTUALIZED SERVER DATA CENTER ENVIRONMENT - A data center includes a physical host machine operating a virtualized entity and a network switch having a physical port connected to the physical host machine. To configure the network switch, the network switch has a management module that acquires information about the virtualized entity operating on the physical host machine. The network switch associates the acquired information about the virtualized entity with the physical port, assigns the virtualized entity to a group associated with a traffic-handling policy, and processes packet traffic from the virtualized entity in accordance with the traffic-handling policy. The virtualized entity can be, for example, a virtual machine or a multi-queue network input/output adapter operating on the physical host machine. | 02-10-2011 |
20110103389 | METHOD AND APPARATUS FOR SWITCHING TRAFFIC BETWEEN VIRTUAL MACHINES - Systems and methods for switching traffic include a physical machine running source and destination virtual machines (VMs). The source VM issues a data unit addressed to the destination VM. The physical machine has a physical network interface in communication with the VMs. The physical network interface transmits a sub-packet, which includes a partial portion of the data unit, over a network while a majority portion of the data unit remains at the physical machine. A network switch on the network receives the sub-packet transmitted by the physical network interface. The network switch performs one or more OSI Layer 2 through Layer 7 switching functions on the sub-packet and returns that sub-packet to the physical network interface. The physical network interface identifies the data unit stored in the memory in response to the sub-packet returned from the network switch and forwards the identified data unit to the destination VM. | 05-05-2011 |
20120093035 | UNIFIED FABRIC PORT - A method and system for configuring communications over a physical communication link connected between a physical port of a network switch and a physical port of a physical network interface on an end station. The communication link between the physical port of the network switch and the physical port of the physical network interface is logically partitioned into a number of channels of communication. For each channel, a channel profile is generated that defines properties of that channel. The physical network interface is instructed to self-configure such that the physical network interface is able to communicate with the network switch over each channel in accordance with the channel profile defined for that channel. | 04-19-2012 |
20120209940 | METHOD FOR SWITCHING TRAFFIC BETWEEN VIRTUAL MACHINES - Methods for switching traffic include a physical machine running source and destination virtual machines (VMs). The source VM issues a data unit addressed to the destination VM. The physical machine has a physical network interface in communication with the VMs. The physical network interface transmits a sub-packet, which includes a partial portion of the data unit, over a network while a majority portion of the data unit remains at the physical machine. A network switch on the network receives the sub-packet transmitted by the physical network interface. The network switch performs one or more OSI Layer 2 through Layer 7 switching functions on the sub-packet and returns that sub-packet to the physical network interface. The physical network interface identifies the data unit stored in the memory in response to the sub-packet returned from the network switch and forwards the identified data unit to the destination VM. | 08-16-2012 |
20120287785 | DATA TRAFFIC HANDLING IN A DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - A switching network includes an upper tier having a master switch and a lower tier including a plurality of lower tier entities. The master switch, which has a plurality of ports each coupled to a respective lower tier entity, implements on each of the ports a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier entity coupled to that port. Data traffic communicated between the master switch and RPIs is queued within virtual ports that correspond to the RPIs with which the data traffic is communicated. The master switch applies data handling to the data traffic in accordance with a control policy based at least upon the virtual port in which the data traffic is queued, such that the master switch applies different policies to data traffic queued to two virtual ports on the same port of the master switch. | 11-15-2012 |
20120287786 | PRIORITY BASED FLOW CONTROL IN A DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - A switching network includes an upper tier and a lower tier including a plurality of lower tier entities. A master switch in the upper tier, which has a plurality of ports each coupled to a respective lower tier entity, implements on each of the ports a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier entity coupled to that port. Data traffic communicated between the master switch and RPIs is queued within virtual ports that correspond to the RPIs on lower tier entities with which the data traffic is communicated. The master switch enforces priority-based flow control (PFC) on data traffic of a given virtual port by transmitting, to a lower tier entity on which a corresponding RPI resides, a PFC data frame specifying priorities for at least two different classes of data traffic communicated by the particular RPI. | 11-15-2012 |
20120287787 | PRIORITY BASED FLOW CONTROL IN A DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - A switching network includes an upper tier and a lower tier including a plurality of lower tier entities. A master switch in the upper tier, which has a plurality of ports each coupled to a respective lower tier entity, implements on each of the ports a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier entity coupled to that port. Data traffic communicated between the master switch and RPIs is queued within virtual ports that correspond to the RPIs on lower tier entities with which the data traffic is communicated. The master switch enforces priority-based flow control (PFC) on data traffic of a given virtual port by transmitting, to a lower tier entity on which a corresponding RPI resides, a PFC data frame specifying priorities for at least two different classes of data traffic communicated by the particular RPI. | 11-15-2012 |
20120287939 | DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - A switching network includes an upper tier including a master switch and a lower tier including a plurality of lower tier entities. The master switch includes a plurality of ports each coupled to a respective one of the plurality of lower tier entities. Each of the plurality of ports includes a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier entity coupled to that port. Each of the plurality of ports also includes a receive interface that, responsive to receipt of data traffic from a particular lower tier entity among the plurality of lower tier entities, queues the data traffic to the virtual port among the plurality of virtual ports that corresponds to the RPI on the particular lower tier entity that was the source of the data traffic. The master switch further includes a switch controller that switches data traffic from the virtual port to an egress port among the plurality of ports from which the data traffic is forwarded. | 11-15-2012 |
20120291034 | TECHNIQUES FOR EXECUTING THREADS IN A COMPUTING ENVIRONMENT - A technique for executing normally interruptible threads of a process in a non-preemptive manner includes in response to a first entry associated with a first message for a first thread reaching a head of a run queue, receiving, by the first thread, a first wake-up signal. In response to receiving the wake-up signal, the first thread waits for a global lock. In response to the first thread receiving the global lock, the first thread retrieves the first message from an associated message queue and processes the retrieved first message. In response to completing the processing of the first message, the first thread transmits a second wake-up signal to a second thread whose associated entry is next in the run queue. Finally, following the transmitting of the second wake-up signal, the first thread releases the global lock. | 11-15-2012 |
20120307684 | METHOD FOR PROVIDING LOCATION INDEPENDENT DYNAMIC PORT MIRRORING ON DISTRIBUTED VIRTUAL SWITCHES - A method for providing location independent dynamic port mirroring on distributed virtual switches is disclosed. A controller is provided to configure one or more virtual switches within a group of physical machines to appear as a set of distributed virtual switches. In response to the receipt of a data packet at a port of a physical machine, a determination is made whether or not the port has a monitor port located on the physical machine. If the port has a monitor port located on the same physical machine, a copy of the data packet is sent to the monitor port of the physical machine. If the port has a monitor port located on a different physical machine, a copy of the data packet along with an identification (ID) of the port and an ID of the monitor port are encapsulated, and the encapsulated information are sent to a controller. | 12-06-2012 |
20120320749 | DATA TRAFFIC HANDLING IN A DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - A switching network includes an upper tier having a master switch and a lower tier including a plurality of lower tier entities. The master switch, which has a plurality of ports each coupled to a respective lower tier entity, implements on each of the ports a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier entity coupled to that port. Data traffic communicated between the master switch and RPIs is queued within virtual ports that correspond to the RPIs with which the data traffic is communicated. The master switch applies data handling to the data traffic in accordance with a control policy based at least upon the virtual port in which the data traffic is queued, such that the master switch applies different policies to data traffic queued to two virtual ports on the same port of the master switch. | 12-20-2012 |
20120324460 | Thread Execution in a Computing Environment - A technique for executing normally interruptible threads of a process in a non-preemptive manner includes in response to a first entry associated with a first message for a first thread reaching a head of a run queue, receiving, by the first thread, a first wake-up signal. In response to receiving the wake-up signal, the first thread waits for a global lock. In response to the first thread receiving the global lock, the first thread retrieves the first message from an associated message queue and processes the retrieved first message. In response to completing the processing of the first message, the first thread transmits a second wake-up signal to a second thread whose associated entry is next in the run queue. Finally, following the transmitting of the second wake-up signal, the first thread releases the global lock. | 12-20-2012 |
20130022050 | DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - A switching network includes an upper tier including a master switch and a lower tier including a plurality of lower tier entities. The master switch includes a plurality of ports each coupled to a respective one of the plurality of lower tier entities. Each port includes a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier entity coupled to that port. Each port also includes a receive interface that, responsive to data traffic from a particular lower tier entity, queues the data traffic to the virtual port that corresponds to the RPI on the particular lower tier entity that was the source of the data traffic. The master switch further includes a switch controller that switches data traffic from the virtual port to an egress port from which the data traffic is forwarded. | 01-24-2013 |
20130064066 | UPDATING A SWITCH SOFTWARE IMAGE IN A DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK - A switching network has a plurality of switches including at least a switch and a managing master switch. At the managing master switch, a first capability vector (CV) is received from the switch. The managing master switch determines whether the first CV is compatible with at least a second CV in a network membership data structure that records CVs of multiple switches in the switching network. In response to detecting an incompatibility, the managing master switch initiates an image update to an image of the switch. In response to a failure of the image update at the switch, the switch boots utilizing a mini-DC module that reestablishes communication between the switch with the managing master switch and retries the image update. | 03-14-2013 |
20130064067 | HIGH AVAILABILITY DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - In a switching network, each of a plurality of lower tier entities is coupled to each of multiple master switches at an upper tier by a respective one of multiple links. At each of the multiple master switches, a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier are implemented on each of a plurality of ports. Each of the plurality of lower tier entities implements a respective egress port mapping indicating which of its plurality of RPIs transmits egress data traffic through each of its multiple links to the multiple master switches. In response to failure of one of the multiple links coupling a particular lower tier entity to a particular master switch, the particular lower tier entity updates its egress port mapping to redirect egress data traffic to another of the multiple master switches without packet dropping. | 03-14-2013 |
20130064068 | HIGH AVAILABILITY DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK ARCHITECTURE - In a switching network, each of a plurality of lower tier entities is coupled to each of multiple master switches at an upper tier by a respective one of multiple links. At each of the multiple master switches, a plurality of virtual ports each corresponding to a respective one of a plurality of remote physical interfaces (RPIs) at the lower tier are implemented on each of a plurality of ports. Each of the plurality of lower tier entities implements a respective egress port mapping indicating which of its plurality of RPIs transmits egress data traffic through each of its multiple links to the multiple master switches. In response to failure of one of the multiple links coupling a particular lower tier entity to a particular master switch, the particular lower tier entity updates its egress port mapping to redirect egress data traffic to another of the multiple master switches without packet dropping. | 03-14-2013 |
20130067049 | UPDATING A SWITCH SOFTWARE IMAGE IN A DISTRIBUTED FABRIC PROTOCOL (DFP) SWITCHING NETWORK - A switching network has a plurality of switches including at least a switch and a managing master switch. At the managing master switch, a first capability vector (CV) is received from the switch. The managing master switch determines whether the first CV is compatible with at least a second CV in a network membership data structure that records CVs of multiple switches in the switching network. In response to detecting an incompatibility, the managing master switch initiates an image update to an image of the switch. In response to a failure of the image update at the switch, the switch boots utilizing a mini-DC module that reestablishes communication between the switch with the managing master switch and retries the image update. | 03-14-2013 |
20130088959 | CREDIT-BASED NETWORK CONGESTION MANAGEMENT - A switching network includes first, second and third switches coupled for communication, such that the first and third switches communicate data traffic via the second switch. The first switch is operable to request transmission credits from the third switch, receive the transmission credits from the third switch and perform transmission of data traffic in reference to the transmission credits. The third switch is operable to receive the request for transmission credits from the first switch, generate the transmission credits and transmit the transmission credits to the first switch via the second switch. The second switch is operable to modify the transmission credits transmitted by the third switch prior to receipt of the transmission credits at the first switch. | 04-11-2013 |
20130088969 | NETWORK TRAFFIC DISTRIBUTION - A switch for a switching network includes a plurality of ports for communicating data traffic and a switch controller that controls switching between the plurality of ports. The switch controller selects a forwarding path for the data traffic based on at least topological congestion information for the switching network. In a preferred embodiment, the topological congestion information includes sFlow topological congestion information and the switch controller includes an sFlow client that receives the sFlow topological congestion information from an sFlow controller in the switching network. | 04-11-2013 |
20130089101 | CREDIT-BASED NETWORK CONGESTION MANAGEMENT - A switching network includes first, second and third switches coupled for communication, such that the first and third switches communicate data traffic via the second switch. The first switch is operable to request transmission credits from the third switch, receive the transmission credits from the third switch and perform transmission of data traffic in reference to the transmission credits. The third switch is operable to receive the request for transmission credits from the first switch, generate the transmission credits and transmit the transmission credits to the first switch via the second switch. The second switch is operable to modify the transmission credits transmitted by the third switch prior to receipt of the transmission credits at the first switch. | 04-11-2013 |
20130182571 | NETWORK TRAFFIC DISTRIBUTION - A switch for a switching network includes a plurality of ports for communicating data traffic and a switch controller that controls switching between the plurality of ports. The switch controller selects a forwarding path for the data traffic based on at least topological congestion information for the switching network. In a preferred embodiment, the topological congestion information includes sFlow topological congestion information and the switch controller includes an sFlow client that receives the sFlow topological congestion information from an sFlow controller in the switching network. | 07-18-2013 |
20130235735 | DIAGNOSTICS IN A DISTRIBUTED FABRIC SYSTEM - A distributed fabric system has distributed line card (DLC) chassis and scaled-out fabric coupler (SFC) chassis. Each DLC chassis includes a network processor and fabric ports. Each network processor of each DLC chassis includes a fabric interface in communication with the DLC fabric ports of that DLC chassis. Each SFC chassis includes a fabric element and fabric ports. A communication link connects each SFC fabric port to one DLC fabric port. Each communication link includes cell-carrying lanes. Each fabric element of each SFC chassis collects per-lane statistics for each SFC fabric port of that SFC chassis. Each SFC chassis includes program code that obtains the per-lane statistics collected by the fabric element chip of that SFC chassis. A network element includes program code that gathers the per-lane statistics collected by each fabric element of each SFC chassis and integrates the statistics into a topology of the entire distributed fabric system. | 09-12-2013 |
20130235762 | MANAGEMENT OF A DISTRIBUTED FABRIC SYSTEM - A distributed fabric system has distributed line card (DLC) chassis and scaled-out fabric coupler (SFC) chassis. Each DLC includes a network processor and fabric ports. Each network processor of each DLC includes a fabric interface in communication with the fabric ports of that DLC. Each SFC includes at least one fabric element and SFC fabric ports. A fabric communication link connects each SFC fabric port to one DLC fabric port. Each fabric communication link includes cell-carrying lanes. Each fabric element of each SFC detects connectivity between each SFC fabric port of that SFC and one DLC fabric port over a fabric communication link. Each SFC includes program code that reads connectivity matrix from fabric element chips and sends connection information corresponding to the detected connectivity from that SFC to a central agent. A network element includes the central agent, which, when executed, constructs a topology of the distributed fabric system from the connection information sent from each SFC. | 09-12-2013 |
20130235763 | MANAGEMENT OF A DISTRIBUTED FABRIC SYSTEM - A distributed fabric system has distributed line card (DLC) chassis and scaled-out fabric coupler (SFC) chassis. Each DLC includes a network processor and fabric ports. Each network processor includes a fabric interface in communication with the fabric ports of that DLC. Each SFC includes at least one fabric element and SFC fabric ports. A fabric communication link connects each SFC fabric port to one DLC fabric port. Each fabric communication link includes cell-carrying lanes. Each fabric element detects connectivity between each SFC fabric port of that SFC and one DLC fabric port over a fabric communication link. Each SFC reads a connectivity matrix from fabric element chips and sends connection information corresponding to the detected connectivity from that SFC to a central agent. A network element includes the central agent, which, when executed, constructs a topology of the distributed fabric system from the connection information sent from each SFC. | 09-12-2013 |
20130259038 | COMMUNICATION TRANSPORT PROTOCOL FOR DISTRIBUTED INFORMATION TECHNOLOGY ARCHITECTURES - A communication protocol in a layer two (L2) network switch comprises, in response to a service request by a source node, registering the source node for packet communication service. The protocol further comprises forwarding one or more packets from the registered source node to one or more destination nodes. The protocol further comprises receiving packets from one or more destination nodes and forwarding each received packet to a corresponding registered node. | 10-03-2013 |
20130259040 | COMMUNICATION TRANSPORT PROTOCOL FOR DISTRIBUTED INFORMATION TECHNOLOGY ARCHITECTURES - A communication protocol in a layer two (L2) network switch comprises, in response to a service request by a source node, registering the source node for packet communication service. The protocol further comprises forwarding one or more packets from the registered source node to one or more destination nodes. The protocol further comprises receiving packets from one or more destination nodes and forwarding each received packet to a corresponding registered node. | 10-03-2013 |
20140010096 | PORT MIRRORING IN DISTRIBUTED SWITCHING SYSTEMS - Port mirroring in a clustered network may be performed between a local switch and a remote switch. A port in the remote switch may be designated a mirrored port where data traffic passing there through can be copied and sent to a mirror-to-port on the local switch. In a virtual local area network (VLAN) environment, data frames of the copied traffic may include a VLAN header identifying the local switch so that routing of the data frames through the network may direct the data frames for monitoring at the local switch. | 01-09-2014 |
20140056152 | PORT MIRRORING IN DISTRIBUTED SWITCHING SYSTEMS - Port mirroring in a clustered network may be performed between a local switch and a remote switch. A port in the remote switch may be designated a mirrored port where data traffic passing there through can be copied and sent to a mirror-to-port on the local switch. In a virtual local area network (VLAN) environment, data frames of the copied traffic may include a VLAN header identifying the local switch so that routing of the data frames through the network may direct the data frames for monitoring at the local switch. | 02-27-2014 |
20140064105 | DIAGNOSTICS IN A DISTRIBUTED FABRIC SYSTEM - A distributed fabric system has distributed line card (DLC) chassis and scaled-out fabric coupler (SFC) chassis. Each DLC chassis includes a network processor and fabric ports. Each network processor of each DLC chassis includes a fabric interface in communication with the DLC fabric ports of that DLC chassis. Each SFC chassis includes a fabric element and fabric ports. A communication link connects each SFC fabric port to one DLC fabric port. Each communication link includes cell-carrying lanes. Each fabric element of each SFC chassis collects per-lane statistics for each SFC fabric port of that SFC chassis. Each SFC chassis includes program code that obtains the per-lane statistics collected by the fabric element chip of that SFC chassis. A network element includes program code that gathers the per-lane statistics collected by each fabric element of each SFC chassis and integrates the statistics into a topology of the entire distributed fabric system. | 03-06-2014 |
20140079075 | SEGMENTATION AND REASSEMBLY OF NETWORK PACKETS FOR SWITCHED FABRIC NETWORKS - Reassembly of member cells into a packet comprises receiving an incoming member cell of a packet from a switching fabric wherein each member cell comprises a segment of the packet and a header, generating a reassembly key using selected information from the incoming member cell header wherein the selected information is the same for all member cells of the packet, checking a reassembly table in a content addressable memory to find an entry that includes a logic key matching the reassembly key, and using a content index in the found entry and a sequence number of the incoming member cell within the packet, to determine a location offset in a reassembly buffer area for storing the incoming member cell at said location offset in the reassembly buffer area for the packet for reassembly. | 03-20-2014 |
20140079076 | SEGMENTATION AND REASSEMBLY OF NETWORK PACKETS - Reassembly of fragments into a packet comprises receiving an incoming fragment of a packet from a network wherein each fragment comprises a segment of the packet and a header, generating a reassembly key using selected information from the incoming fragment header wherein the selected information is the same for all fragments of the packet, checking a reassembly table in a content addressable memory to find an entry that includes a logic key matching the reassembly key, and using a content index in the found entry and a sequence number of the incoming fragment within the packet, to determine a location offset in a reassembly buffer area for storing the incoming fragment at said location offset in the reassembly buffer area for the packet for reassembly. | 03-20-2014 |
20150078389 | SEGMENTATION AND REASSEMBLY OF NETWORK PACKETS FOR SWITCHED FABRIC NETWORKS - Reassembly of member cells into a packet comprises receiving an incoming member cell of a packet from a switching fabric wherein each member cell comprises a segment of the packet and a header, generating a reassembly key using selected information from the incoming member cell header wherein the selected information is the same for all member cells of the packet, checking a reassembly table in a content addressable memory to find an entry that includes a logic key matching the reassembly key, and using a content index in the found entry and a sequence number of the incoming member cell within the packet, to determine a location offset in a reassembly buffer area for storing the incoming member cell at said location offset in the reassembly buffer area for the packet for reassembly. | 03-19-2015 |