Patent application number | Description | Published |
20090210382 | METHOD FOR PRIORITY SEARCH USING A TCAM - Methods and apparatus are disclosed for searching prioritized content using a ternary content-addressable memory (TCAM). Data strings and corresponding priorities are stored at indexed storage locations in a TCAM. The priorities for the corresponding data strings are unsorted with respect to the indices of the indexed storage locations. The TCAM is searched using input data including an input string, an input mask, an input priority and a priority mask. The number of searches is no more than the number of possible distinct priorities—in some embodiments no more than the logarithm of the number of possible priorities. A match index is output corresponding to a storage location that has a matching string selected according to the input mask and also has an optimal priority. | 08-20-2009 |
20100058336 | ASSIGNMENT, AT LEAST IN PART, OF AT LEAST ONE VIRTUAL MACHINE TO AT LEAST ONE PACKET - In an embodiment, an apparatus is provided that may include circuitry to be comprised in a platform. The circuitry may be to determine, at least in part, at least one virtual machine (VM) resident in the platform to be assigned to process, at least in part, at least one packet. The circuitry may determine, at least in part, the at least one VM based at least in part upon (1) comparison, at least in part, of at least one destination of the at least one packet to at least one predetermined destination, and (2) at least one value based at least in part upon both at least one port associated with the at least one packet and at least one source of the at least one packet. Of course, many alternatives, variations, and modifications are possible without departing from this embodiment. | 03-04-2010 |
20120002680 | INTERRUPTION, AT LEAST IN PART, OF FRAME TRANSMISSION - An embodiment may include circuitry to permit interruption, at least in part, of a first frame from a sender to an intended recipient in favor of transmitting, at least in part, a payload of a second frame from the sender to the intended recipient, and/or processing, at least in part, one or more incoming flow control notifications. The payload may be transmitted, at least in part, to the intended recipient in one or more frame fragments. Many modifications, variations, and alternatives are possible without departing from this embodiment. | 01-05-2012 |
20120254480 | FACILITATING, AT LEAST IN PART, BY CIRCUITRY, ACCESSING OF AT LEAST ONE CONTROLLER COMMAND INTERFACE - An embodiment may include circuitry to facilitate, at least in part, a first network interface controller (NIC) in a client to be capable of accessing, via a second NIC in a server that is remote from the client and in a manner that is independent of an operating system environment in the server, at least one command interface of another controller of the server. The command interface may include at least one controller command queue. Such accessing may include writing at least one queue element to the at least one command queue to command the another controller to perform at least one operation associated with the another controller. The another controller may perform the at least one operation in response, at least in part, to the at least one queue element. Many alternatives, variations, and modifications are possible. | 10-04-2012 |
20130031568 | Virtualizable and Forward-Compatible Hardware-Software Interface - Methods and apparatus are disclosed for virtualizable, forward-compatible hardware-software interfaces. Embodiments may be used in a driver whether it is a physical driver or a virtual driver. Commands are queued from the driver and fetched to the device. An actions table is accessed to determine if drivers are permitted to perform commands. Events are queued for the drivers responsive to commands. If drivers are not permitted to perform a command, device firmware may forward the command to a privileged driver to perform the required command. If a driver is only permitted to perform a command with assistance the command is forwarded for corrections and execution. If a command is to be dropped, a completion event may be queued as if the command had executed. Drivers may have no indication of which actions were taken. The actions table may be changed for hardware/software modifications or dynamically according to configuration changes. | 01-31-2013 |
20130246828 | AT LEAST ONE MESSAGE TO ANNOUNCE ENTRY INTO RELATIVELY LOWER POWER STATE - An embodiment may include circuitry that may be capable of performing operations that may include generating, at least in part, at least one message to announce that at least one network node (1) is requesting, at least in part, that one or more transmissions to the at least one network node be postponed, at least in part, and/or (2) is entering, at least in part after issuance of the at least one message, a relatively lower power state relative to a relatively higher power state. Additionally or alternatively, the operations may include, in response, at least in part, to the at least one message, postponing, at least in part, at least one intermediate node at least one transmission (received by the at least one intermediate node) to the at least one network node. Many alternatives, variations, and/or modifications are possible without departing from this embodiment. | 09-19-2013 |
20130262711 | TECHNIQUES FOR USING AN ASSIGNED SWITCH IDENTIFICATION AT AN INPUT/OUTPUT DEVICE - Examples are disclosed for assigning a switch identification to data received at an input/output device coupled to a host device. In some examples, the data may be associated with a virtual station interface (VSI) for a virtual machine implemented at the host device. For these examples, a switch identification may be assigned to the data based on an assigned port identifier, identification information for the data, or an originator of at least a portion of the identification information. The assigned switch identification may then be used as part of a lookup table to determine one or more actions for processing the data at the input/output device. Other examples are described and claimed. | 10-03-2013 |
20130262868 | SHARED BUFFERS FOR PROCESSING ELEMENTS ON A NETWORK DEVICE - Examples are disclosed for exchanging a key between an input/output device for network device and a first processing element operating on the network device. Data having a destination associated with the first processing element may be received by the input/output device. The exchanged key may be used to encrypt the received data. The encrypted data may then be sent to a buffer maintained at least in part in a memory for the network device. The memory may be arranged to enable sharing of the buffer with at least a second processing element operating on the network device. Examples are also disclosed for the processing element to receive an indication of the storing of the encrypted data in the buffer. The processing element may then obtain the encrypted data from the buffer and decrypt the data using the exchanged key. | 10-03-2013 |
20140059266 | METHODS AND APPARATUS FOR SHARING A NETWORK INTERFACE CONTROLLER - Methods, apparatus, and systems for enhancing communication between compute resources and networks in a micro-server environment. Micro-server modules configured to be installed in a server chassis include a plurality of processor subsystems coupled in communication to a shared Network Interface Controller (NIC) via PCIe links. The shared NIC includes at least one Ethernet port and a PCIe block including a shared PCIe interface having a first number of lanes. The PCIe lines between the processor sub-systems and the shared PCIe interface employ a number of lanes that is less than the first number of lanes, and during operation of the micro-server module, the shared NIC is configured to enable each processor sub-system to access the at least one Ethernet port using the PCIe link between that processor sub-system and the shared PCIe block on the shared NIC. | 02-27-2014 |
20140112346 | SYSTEM AND METHOD PROVIDING FORWARD COMPATIBILITY BETWEEN A DRIVER MODULE AND A NETWORK INTERFACE - Generally, this disclosure provides systems and methods for providing forward compatibility between a driver module and one or more present or future versions of a network interface. The system may include a network interface configured to transfer data between a host system and a network; and a programmable circuit module associated with the network interface, the programmable circuit module configured to provide compatibility between the network interface and a driver module associated with the host system, wherein the driver module includes a first set of capabilities and the network interface includes a second set of capabilities. | 04-24-2014 |
20140136646 | FACILITATING, AT LEAST IN PART, BY CIRCUITRY, ACCESSING OF AT LEAST ONE CONTROLLER COMMAND INTERFACE - An embodiment may include circuitry to facilitate, at least in part, a first network interface controller (NIC) in a client to be capable of accessing, via a second NIC in a server that is remote from the client and in a manner that is independent of an operating system environment in the server, at least one command interface of another controller of the server. The command interface may include at least one controller command queue. Such accessing may include writing at least one queue element to the at least one command queue to command the another controller to perform at least one operation associated with the another controller. The another controller may perform the at least one operation in response, at least in part, to the at least one queue element. Many alternatives, variations, and modifications are possible. | 05-15-2014 |
20140153574 | NOTIFICATION BY NETWORK ELEMENT OF PACKET DROPS - Methods and apparatus for implementing notification by network elements of packet drops. In response to determining a packet is to be dropped, a network element such as a switch or router determines the source of the packet and returns a dropped packet notification message to the source. Upon receipt of notification, networking software or embedded hardware on the source causes the dropped packet to be retransmitted. The notification may also be sent from the network element to the destination computer to inform networking software or embedded logic implemented by the destination computer that the packet was dropped and notification to the source has been sent, thus alleviating the destination from needing to send a Selective ACKnowledge (SACK) message to inform the source the packet was not delivered. (Too narrow) | 06-05-2014 |
20140207906 | TECHNIQUES TO COPY AN OPERATING SYSTEM - Techniques are described that can be used to enable a transfer of an operating system from one machine to another. The transfer permits the operating system to be available to the target machine at buffers that are accessible to one or more application or other logic. In some implementations, information related to an operating system migration is stored in a buffer that is accessible to an application that is to use the information and thereby avoids a copy of such information from an intermediate buffer to an application buffer. | 07-24-2014 |
20140301198 | LINK STATUS BUFFER FLOW CONTROL MANAGEMENT - Generally, this disclosure describes techniques for buffer management based on link status. A host platform may include a Baseboard Management Controller (BMC) and a network controller that includes a buffer used by the BMC. When a network controller is in a lower power link state, the BMC may attempt to send data to the link partner which causes the network controller to transition out of the low power state. However, this transition may take longer than the buffer's ability to buffer the incoming flow from the BMC. Accordingly, to avoid the need for larger buffer space, a buffer manager is used to provide flow control management of the buffer based on link status. | 10-09-2014 |
20150049769 | SOCKET MANAGEMENT WITH REDUCED LATENCY PACKET PROCESSING - Generally, this disclosure provides systems, methods and computer readable media for management of sockets and device queues for reduced latency packet processing. The method may include maintaining a unique-list comprising entries identifying device queues and an associated unique socket for each of the device queues, the unique socket selected from a plurality of sockets configured to receive packets; busy-polling the device queues on the unique-list; receiving a packet from one of the plurality of sockets; and updating the unique-list in response to detecting that the received packet was provided by an interrupt processing module. The updating may include identifying a device queue associated with the received packet; identifying a socket associated with the received packet; and if the identified device queue is not on one of the entries on the unique-list, creating a new entry on the unique-list, the new entry comprising the identified device queue and the identified socket. | 02-19-2015 |