Patent application number | Description | Published |
20090254647 | SYSTEM AND METHOD FOR NETWORK INTERFACING - Systems and methods for network interfacing may include a communication data center with a first tier, a second tier and a third tier. The first tier may include a first server with a first single integrated convergent network controller chip. The second server may include a second server with a second single integrated convergent network controller chip. The third tier may include a third server with a third single integrated convergent network controller chip. The second server may be coupled to the first server via a single fabric with a single connector. The third server may be coupled to the second server via the single fabric with the single connector. The respective first, second and third server, each processes a plurality of different traffic types concurrently via the respective first, second and third single integrated convergent network chip over the single fabric that is coupled to the single connector. | 10-08-2009 |
20100250783 | SYSTEM AND METHOD FOR TCP/IP OFFLOAD INDEPENDENT OF BANDWIDTH DELAY PRODUCT - A network interface device may include an offload engine that receives control of state information while a particular connection is offloaded. Control of the state information for the particular connection may be split between the network interface device and a host. The at least one connection variables may be updated and provided to the host. | 09-30-2010 |
20110040891 | System and Method for TCP Offload - A system for processing packets is disclosed and may including a network interface card (NIC). The NIC may include a TCP enabled Ethernet controller (TEEC). The TEEC may include an internal elastic buffer. The TEEC may process received incoming TCP packets once and may temporarily buffer at least a portion of the incoming TCP packets in the internal elastic buffer. The processing may occur without reassembly or retransmission. The internal elastic buffer may include a receive internal elastic buffer and a transmit internal elastic buffer. The receive internal elastic buffer may temporarily buffer at least a portion of the received incoming TCP packets. The transmit internal elastic buffer may temporarily buffer at least a portion of TCP packets to be transmitted. The TEEC may place at least a portion of the received incoming TCP packets data into at least a portion of a host memory. | 02-17-2011 |
20110185076 | System and Method for Network Interfacing - Systems and methods for network interfacing may include a communication data center with a first tier, a second tier and a third tier. The first tier may include a first server with a first single integrated convergent network controller chip. The second server may include a second server with a second single integrated convergent network controller chip. The third tier may include a third server with a third single integrated convergent network controller chip. The second server may be coupled to the first server via a single fabric with a single connector. The third server may be coupled to the second server via the single fabric with the single connector. The respective first, second and third server, each processes a plurality of different traffic types concurrently via the respective first, second and third single integrated convergent network chip over the single fabric that is coupled to the single connector. | 07-28-2011 |
20110202702 | METHOD AND SYSTEM FOR PCI HYBRID FUNCTION - A single root I/O virtualization (SR-IOV) capable peripheral component interconnect (PCI) device may be operable to configure resources to transparently operate in a non-SR-IOV environment utilizing a physical function (PF) and one or more hybrid functions (HFs). In instances when the SR-IOV capable PCI device is operating in a SR-IOV environment, the SR-IOV capable PCI device may expose a VF configuration space in the hybrid function to an associated PCI driver for configuring the hybrid function as a virtual function. In instances when the SR-IOV capable PCI device is operating in a non-SR-IOV environment, the SR-IOV capable PCI device may hide a SR-IOV extended header in the physical function from the associated PCI driver and expose a PF configuration space in the hybrid function to the associated PCI driver for configuring the hybrid function as a physical function. | 08-18-2011 |
20110213863 | METHOD AND SYSTEM FOR NIC-CENTRIC HYPER-CHANNEL DISTRIBUTED NETWORK MANAGEMENT - A server comprising a network interface controller (NIC) may be operable to determine network components and/or resources that may be utilized to create, modify and/or terminate a hyper-channel in a network fabric. A configuration for creating, modifying and/or terminating the hyper-channel may be initiated by the NIC based on knowledge of the network fabric up to the edge of network and each of the switches along a path of the hyper-channel may be instructed by the NIC to configure the hyper-channel. The configuration may also be initiated by the NIC based on knowledge of the network fabric up to a first hop and each of the switches along a path of the hyper-channel may then be allowed to configure the hyper-channel based on knowledge of the network fabric surrounding each of the switches. | 09-01-2011 |
20110246662 | SYSTEM AND METHOD FOR TCP OFFLOAD - Aspects of the invention may comprise receiving an incoming TCP packet at a TEEC and processing at least a portion of the incoming packet once by the TEEC without having to do any reassembly and/or retransmission by the TEEC. At least a portion of the incoming TCP packet may be buffered in at least one internal elastic buffer of the TEEC. The internal elastic buffer may comprise a receive internal elastic buffer and/or a transmit internal elastic buffer. Accordingly, at least a portion of the incoming TCP packet may be buffered in the receive internal elastic buffer. At least a portion of the processed incoming packet may be placed in a portion of a host memory for processing by a host processor or CPU. Furthermore, at least a portion of the processed incoming TCP packet may be DMA transferred to a portion of the host memory. | 10-06-2011 |
20130124709 | METHOD AND SYSTEM FOR NIC-CENTRIC HYPER-CHANNEL DISTRIBUTED NETWORK MANAGEMENT - Various embodiments for hyper-channel management are disclosed. One such method is implemented by a network interface controller (NIC) residing in a server. This method involves determining a network component that is operable to create, modify, terminate, or any combination thereof, a hyper-channel. The hyper-channel corresponds to a logical abstraction of network resources associated with the server. The method also involves using the determined network component to manage the hyper-channel. | 05-16-2013 |
Patent application number | Description | Published |
20100057941 | Method and System for Providing Smart Offload and Upload - Provided is a system and method for performing smart offloads between computer applications module and a network interfacing device within a data communications system. The method includes receiving data requests from the computer applications module and determining whether the received data requests require offloading. The received requests are forwarded along a first data path to the network interfacing device if offloading is required. If offloading is not required, the received data requests are forwarded along a secondary data path to a host protocol stack for processing. Next, the protocol processing is performed and the processed data requests are forwarded to the network interfacing device. | 03-04-2010 |
20100115118 | SYSTEM AND METHOD FOR MANAGING MULTIPLE STACK ENVIRONMENTS - Systems and methods that manage multiple stack environments are provided. In one example, a system may include, for example, a first protocol processing stack, a second protocol processing stack and a mapper. The mapper may be coupled to the first protocol processing stack and to the second protocol processing stack. A first port number may be associated with the first protocol processing stack and a second port number may be associated with the second protocol processing stack. The mapper may store a correspondence between the first port number and the second port number. | 05-06-2010 |
20100121995 | SYSTEM AND METHOD FOR SUPPORTING TCP OUT-OF-ORDER RECEIVE DATA USING GENERIC BUFFER - A method and system for handling received out-of-order network data using generic buffers for non-posting TCP applications is disclosed. When incoming out-of-order data is received and there is no application buffer posted, a TCP data placement may notify a TCP reassembler to terminate a current generic buffer, allocate a new current generic buffer, and DMA the incoming data into the new current generic buffer. The TCP data placement may notify the TCP reassembler the starting TCP sequence number and the length of the new current generic buffer. Moreover, the TCP data placement may add entries into a TCP out-of-order table when the incoming data creates a new disjoint area. The TCP data placement may adjust an existing disjoint area to reflect any updates. When a TCP application allocates or posts a buffer, then the TCP reassembler may copy data from a linked list of generic buffers into posted buffers. | 05-13-2010 |
20100262859 | SYSTEM AND METHOD FOR FAULT TOLERANT TCP OFFLOAD - Systems and methods that provide fault tolerant transmission control protocol (TCP) offloading are provided. In one example, a method that provides fault tolerant TCP offloading is provided. The method may include one or more of the following steps: receiving TCP segment via a TCP offload engine (TOE); calculating a TCP sequence number; writing a receive sequence record based upon at least the calculated TCP sequence number to a TCP sequence update queue in a host; and updating a first host variable with a value from the written receive sequence record. | 10-14-2010 |
20100306521 | METHOD AND SYSTEM FOR ISCSI BOOT IN WHICH AN ISCSI CLIENT LOADS BOOT CODE FROM A HOST BUS ADAPTER AND/OR NETWORK INTERFACE CARD - Certain aspects of a method for iSCSI boot may include loading boot BIOS code from a host bus adapter or a network interface controller (NIC) by an iSCSI client device. A connection may be established to an iSCSI target by the iSCSI client device after loading the boot BIOS code. The boot BIOS code may be chained to at least one interrupt handler over iSCSI protocol. An operating system may be remotely booted from the iSCSI target by the iSCSI client device based on chaining the interrupt handler. An Internet protocol (IP) address and/or location of the iSCSI target may be received. At least one iSCSI connection may be initiated to the iSCSI target based on chaining at least one interrupt handler. The iSCSI target may be booted in real mode if at least one master boot record is located in the memory. | 12-02-2010 |
20110185089 | Method and System for Supporting Hardware Acceleration for iSCSI Read and Write Operations and iSCSI Chimney - Certain aspects of a method and system for supporting hardware acceleration for iSCSI read and write operations via a TCP offload engine may comprise pre-registering at least one buffer with hardware. An iSCSI command may be received from an initiator. An initiator test tag value, a data sequence value and/or a buffer offset value of an iSCSI buffer may be compared with the pre-registered buffer. Data may be fetched from the pre-registered buffer based on comparing the initiator test tag value, the data sequence value and/or the buffer offset value of the iSCSI buffer with the pre-registered buffer. The fetched data may be zero copied from the pre-registered buffer to the initiator. | 07-28-2011 |
20110307577 | METHOD AND SYSTEM FOR TRANSMIT SCHEDULING FOR MULTI-LAYER NETWORK INTERFACE CONTROLLER (NIC) OPERATION - A method for processing network data includes collecting by a network interface controller (NIC), a plurality of transmit (TX) buffer indicators into a plurality of priority lists of connections. Each of the plurality of TX buffer indicators identifies transmit-ready data located externally to the NIC and not previously received by the NIC. One or more of the plurality of TX buffer indicators may be selected. The identified transmit-ready data may be retrieved into the NIC based on the selected one or more of the plurality of TX buffer indicators. At least a portion of the identified transmit-ready data may be transmitted. Each of the plurality of priority lists may be generated based on a particular connection priority characteristic and a particular connection type. The identified transmit-ready data may be associated with the same connection priority characteristic and the same connection type. | 12-15-2011 |
20110314171 | SYSTEM AND METHOD FOR PROVIDING POOLING OR DYNAMIC ALLOCATION OF CONNECTION CONTEXT DATA - A method for processing of packetized data is disclosed and includes allocating a plurality of partitions of a single context memory for handling data for a corresponding plurality of network protocol connections. Data for at least one of the plurality of network protocol connections may be processed utilizing a corresponding at least one of the plurality of partitions of the single context memory. The at least one of the plurality of partitions of the single context memory may be de-allocated, when the corresponding at least one of the plurality of network protocol connections is terminated. The data for the at least one of the plurality of network protocol connections may be received. The data may be associated with a single network protocol or with a plurality of network protocols. The data for the at least one of the plurality of network protocol connections includes context data. | 12-22-2011 |