Patent application number | Description | Published |
20080235430 | Creation and Management of Routing Table for PCI Bus Address Based Routing with Integrated DID - A method is provided for creating and managing tables for routing packets through an environment that includes multiple hosts and shared PCI switches and adapters. A Destination Identification (DID) field in the PBA is appended to a transaction packet dispatched through the PCI switches, wherein a particular DID is associated with a particular host or system image, and thus identifies the physical or virtual end point of its packet. In one embodiment, packets are routed through PCI switches in a distributed computer system comprising multiple root nodes, wherein each root node includes one or more hosts. The embodiment includes the step of creating a table or like data structure in a specified one of the switches. When a particular host of one of the root nodes becomes connected to the specified switch, a PCI Configuration Master (PCM), residing in one of the root nodes, is operated to enter a destination identifier or DID into the table. The DID is then appended as an address component, to packets directed through the specified switch from the particular host to one of the adapters. The destination identifier is also used to determine that a PCI packet, routed through the specified switch from one of the adapters, is intended for the particular root node. | 09-25-2008 |
20080235431 | Method Using a Master Node to Control I/O Fabric Configuration in a Multi-Host Environment - A method is directed to use of a master root node, in a distributed computer system provided with multiple root nodes, to control the configuration of routings through an I/O switched-fabric. One of the root nodes is designated as the master root node or PCI Configuration Manager (PCM), and is operable to carry out the configuration while each of the other root nodes remains in a quiescent or inactive state. In one useful embodiment pertaining to a system of the above type, that includes multiple root nodes, PCI switches, and PCI adapters available for sharing by different root nodes, a method is provided wherein the master root node is operated to configure routings through the PCI switches. Respective routings are configured between respective root nodes and the PCI adapters, wherein each of the configured routings corresponds to only one of the root nodes. A particular root node is enabled to access each of the PCI adapters that are included in any configured routing that corresponds to the particular root node. At the same time, the master root node writes into a particular root node only the configured routings that correspond to the particular root node. Thus, the particular root node is prevented from accessing an adapter that is not included in its corresponding routings. | 09-25-2008 |
20080235785 | Method, Apparatus, and Computer Program Product for Routing Packets Utilizing a Unique Identifier, Included within a Standard Address, that Identifies the Destination Host Computer System - A computer-implemented method, apparatus, and computer program product are disclosed in a data processing environment that includes host computer systems that are coupled to adapters utilizing a switched fabric for routing packets between the host computer systems and the adapters. A unique destination identifier is assigned to one of the host computer systems. A portion of a standard format packet destination address is selected. Within a particular packet, the portion is set equal to the unique identifier that is assigned to the host computer system. The particular packet is then routed through the fabric to the host computer system using the unique destination identifier. | 09-25-2008 |
20080270853 | Method of Routing I/O Adapter Error Messages in a Multi-Host Environment - A method and apparatus is provided for routing error messages in a distributed computer system comprising multiple root nodes, and further comprising one or more PCI switches and one or more I/O adapters, wherein each root node includes one or more system images. In one useful embodiment, a method is provided for routing I/O error messages to root nodes respectively associated with the errors contained in the messages. The method includes detecting occurrence of an error at a specified one of the adapters, wherein the error affects one of the system images, and generating an error message at the specified adapter. The method further comprises routing the error message from the specified adapter to the particular root node that includes the affected system image. The error message is then selectively processed at the particular root node, in order to identify the affected system image. Usefully, the step of routing the error message includes using a bus/device/function number associated with the error, together with a routing table located in one of the PCI switches, to route the error message to the correct root node and system image. | 10-30-2008 |
20080307116 | Routing Mechanism in PCI Multi-Host Topologies Using Destination ID Field - Method and system for address routing in a distributed computing system, such as a distributed computing system that uses PCI Express protocol to communicate over an I/O fabric. A destination identifier is provided to identify a physical or virtual host or end point. When a physical or virtual host or end point receives a PCI data packet it compares a list of source identifiers with destination identifiers to determine if a source identifier included in the transaction packet is associated with a destination identifier included in the transaction packet to determine if the transaction packet has a valid association. If the transaction packet has a valid association, it is routed to the target device. The present invention enables each host that attaches to PCI bridges or switches and shares a set of common PCI devices to have its own PCI 64-bit address space and enables the routing of PCI transaction packets between multiple hosts and adapters, through a PCI switched-fabric bus using a destination identifier. | 12-11-2008 |
20090037941 | MULTIPLE PARTITION ADJUNCT INSTANCES INTERFACING MULTIPLE LOGICAL PARTITIONS TO A SELF-VIRTUALIZING INPUT/OUTPUT DEVICE - Multiple logical partitions are provided access to a self-virtualizing input/output device of a data processing system via multiple dedicated partition adjunct instances. Access is established by: interfacing each logical partition to one or more associated partition adjunct instances, each partition adjunct instance coupling its associated logical partition to one of a virtual function or a queue pair of the self-virtualizing input/output device, and each partition adjunct instance being a separate dispatchable state and being created employing virtual address space donated from the respective logical partition or a hypervisor of the data processing system, and each partition adjunct instance including a device driver for the virtual function or queue pair of the self-virtualizing input/output device; and providing each logical partition with at least one virtual input/output which is interfaced through the logical partition's respective partition adjunct instance(s) to a virtual function or queue pair of the self-virtualizing input/output device. | 02-05-2009 |
20090052345 | Method and Apparatus for an Adapter in a Network Device to Discover its Adapter Name in a Network System - A network system supports multiple network communication protocols. In one embodiment, network device driver software provides a “Fibre Channel over Ethernet” communication capability and methodology. Device driver software manages a Fibre Channel to Ethernet and Ethernet to Fibre Channel address translation in real time for data packet communications in the network system. Different embodiments of the disclosed network system include multiple name servers and network device driver software that together provide multiple adapter name discovery methodologies. In one embodiment, the adapter name discovery methodologies include port name discovery and adapter attributes discovery. | 02-26-2009 |
20090052346 | Method and Apparatus for Enabling an Adapter in a Network Device to Discover the Name of Another Adapter of Another Network Device in a Network System - A network system supports multiple network communication protocols. In one embodiment, network device driver software provides a “Fibre Channel over Ethernet” communication capability and methodology. Device driver software manages a Fibre Channel to Ethernet and Ethernet to Fibre Channel address translation in real time for data packet communications in the network system. Different embodiments of the disclosed network system include multiple name servers and network device driver software that together provide multiple adapter name discovery methodologies. In one embodiment, the adapter name discovery methodologies include port name discovery and adapter attributes discovery. | 02-26-2009 |
20090052461 | Method and Apparatus for Fibre Channel Over Ethernet Data Packet Translation Via Look up Table Conversion Bridge in a Network System - A network system supports multiple network communication protocols. In one embodiment, network device driver software provides a “Fibre Channel over Ethernet” communication capability and methodology. Device driver software manages a Fibre Channel to Ethernet and Ethernet to Fibre Channel address translation in real time for data packet communications in the network system. Different embodiments of the disclosed network system include multiple name servers and network device driver software that together provide multiple adapter name discovery methodologies. In one embodiment, the adapter name discovery methodologies include port name discovery and adapter attributes discovery. | 02-26-2009 |
20090100204 | Method, Apparatus, and Computer Usable Program Code for Migrating Virtual Adapters from Source Physical Adapters to Destination Physical Adapters - A computer-implemented method, apparatus, and computer usable program code are disclosed for migrating a virtual adapter from a source physical adapter to a destination physical adapter in a data processing system where multiple host computer systems share multiple adapters and communicate with those adapters through a PCI switched-fabric bus. The virtual adapter is first caused to stop processing transactions. All in-flight transactions that are associated with the virtual adapter are then captured. The configuration information that defines the virtual adapter is moved from the source physical adapter to the destination physical adapter. The in-flight transactions are then restored to their original locations on the destination virtual adapter. The virtual adapter is then restarted on the destination physical adapter such that the virtual adapter begins processing transactions. | 04-16-2009 |
20090106771 | System and Method for Providing a Memory Region or Memory Window Access Notification on a System Area Network - A system and method for providing a memory region/memory window (MR/MW) access notification on a system area network are provided. Whenever a previously allocated MR/MW is accessed, such as via a remote direct memory access (RDMA) read/write operation, a notification of the access is generated and written to a queue data structure associated with the MR/MW. In one illustrative embodiment, this queue data structure may be a MR/MW event queue (EQ) data structure that is created and used for all consumer processes and all MR/MWs. In other illustrative embodiments, the EQ is associated with a protection domain. In yet another illustrative embodiment, an event record may be posted to an asynchronous event handler in response to the accessing of the MR/MW. In another illustrative embodiment, a previously posted queue element may be used to generate a completion queue element in response to the accessing of the MR/MW. | 04-23-2009 |
20090119551 | Broadcast of Shared I/O Fabric Error Messages in a Multi-Host Environment to all Affected Root Nodes - A method, mechanism and computer usable medium is provided for distributing I/O fabric errors to the appropriate root nodes in a multi-root environment. The case where the I/O fabric is attached to more than one root node and where each root can potentially share with the other roots the I/O adapter (IOA) resources which are attached to the I/O is addressed. Additionally, a method, mechanism and computer usable medium is provided by which errors detected in an I/O fabric may be routed to all root nodes which may be affected by the error, while not being reported to the root nodes that will not be affected by those errors. In particular, distributed computing system which uses the PCI Express protocol to communicate over the I/O fabric is addressed. | 05-07-2009 |
20090133016 | System and Method for Management of an IOV Adapter Through a Virtual Intermediary in an IOV Management Partition - The system and method address the situation where an input/output (I/O) fabric is shared by more than one logical partition (LPAR) and where each LPAR can share with the other LPARs an I/O adapter (IOA). In particular, each LPAR is assigned its own separate address space to access a virtual function (VF) assigned to it such that each LPAR's perception is that it has its own independent IOA. Each VF may be shared across multiple LPARs. Facilities are provided for management of the shared resources of the IOA via a Physical Function (PF) of the IOA by assignment of that PF to an I/O Virtualization Management Partition (IMP). The code running in the IMP acts as a virtual intermediary to the VFs for fully managing the VF error handling, VF reset, and configuration operations. The IMP also acts as an interface to the PF for accessing common VF functionality. | 05-21-2009 |
20090133028 | SYSTEM AND METHOD FOR MANAGEMENT OF AN IOV ADAPTER THROUGH A VIRTUAL INTERMEDIARY IN A HYPERVISOR WITH FUNCTIONAL MANAGEMENT IN AN IOV MANAGEMENT PARTITION - A system and method which provide a mechanism for an I/O virtualization management partition (IMP) to control the shared functionality of an I/O virtualization (IOV) enabled I/O adapter (IOA) through a physical function (PF) of the IOA while the virtual functions (VFs) are assigned to client partitions for normal I/O operations directly. A hypervisor provides device-independent facilities to the code running in the IMP and client partitions. The IMP may include device specific code without the hypervisor needing to sacrifice its size, robustness, and upgradeability. The hypervisor provides the virtual intermediary functionally for the sharing and control of the IOA's control functions. | 05-21-2009 |
20090141727 | Method and System for Infiniband Over Ethernet by Mapping an Ethernet Media Access Control (MAC) Address to an Infiniband Local Identifier (LID) - A method of providing a converged InfiniBand over Ethernet (IBOE) network. An IBOE switch receives a data packet. If the data packet is received by the IBOE switch from the InfiniBand network, a translation utility generates an Ethernet media access control (MAC) address corresponding to an InfiniBand Local Identifier (LID) of said data packet. The translation utility generates an Ethernet packet and encapsulates an InfiniBand link layer packet before sending the packet. If the data packet is received by the IBOE switch from the Ethernet network, the translation utility removes the Ethernet header from the packet and looks up a LID corresponding to the MAC address. The translation utility calculates an outbound port number and sends the packet. | 06-04-2009 |
20090141728 | METHOD AND SYSTEM FOR PROVIDING VISIBILITY OF ETHERNET COMPONENTS TO A SUBNET MANAGER IN A CONVERGED INFINIBAND OVER ETHERNET NETWORK - A method of providing visibility of Ethernet components to a subnet manager in a converged InfiniBand over Ethernet (IBOE) network. If a port of an IBOE gateway corresponds to one or more InfiniBand devices, the subnet manager sends fabric management packets (FMPs) to discover the InfiniBand network and assigns physical local identifiers (LIDs) to the InfiniBand devices. If a port of the IBOE gateway corresponds to one or more Ethernet devices, the subnet manager sends FMPs to discover the Ethernet network. The subnet manager adds the Ethernet Media Access Control (MAC) addresses of any responding devices to an LID routing table and assigns LIDs to the Ethernet devices. The subnet manager configures one or more virtual Host Channel Adapters (HCAs) corresponding to the one or more Ethernet MAC addresses in the LID routing table. | 06-04-2009 |
20090141734 | METHOD AND SYSTEM FOR A CONVERGED INFINIBAND OVER ETHERNET NETWORK - A method of providing a converged InfiniBand over Ethernet (IBOE) network. An IBOE switch receives a data packet. If the data packet is received by the IBOE switch from the InfiniBand network, a translation utility looks up an Ethernet media access control (MAC) address corresponding to an LID of the packet in a translation table. The translation utility generates an Ethernet packet and encapsulates an InfiniBand link layer packet before sending the packet. If the data packet is received by the IBOE switch from the Ethernet network, the translation utility removes the Ethernet header from the packet and looks up a LID corresponding to the MAC address. The translation utility calculates an outbound port number and sends the packet. | 06-04-2009 |
20090144508 | PCI Express Address Translation Services Invalidation Synchronization with TCE Invalidation - A PCI Express (PCIe) computer system utilizes address translation services to translate virtual addresses from I/O device adaptors to physical addresses of system memory. A combined memory controller and host bridge uses a translation agent to convert the I/O addresses via translation control entries (TCEs) in a TCE table (also known as an address translation and protection table). Some of the I/O device adaptors have address translation caches for local storage of TCEs. The TCE definition includes a new non-cacheable control bit which is set active in the TCE table when the TCE is in the process of being invalidated. The memory controller prevents further caching of the TCE while the non-cacheable control bit is active. A further implementation utilizes a change-in-progress control bit of the TCE to indicate that the TCE is in the process of being changed to allow simultaneous invalidation of the previously TCE information. | 06-04-2009 |
20090144731 | SYSTEM AND METHOD FOR DISTRIBUTION OF RESOURCES FOR AN I/O VIRTUALIZED (IOV) ADAPTER AND MANAGEMENT OF THE ADAPTER THROUGH AN IOV MANAGEMENT PARTITION - The system and method address the situation where an input/output (I/O) fabric is shared by more than one logical partition (LPAR) and where each LPAR can share with the other LPARs an I/O adapter (IOA). In particular, each LPAR is assigned its own separate address space to access a virtual function (VF) assigned to it such that each LPAR's perception is that it has its own independent IOA. Each VF may be shared across multiple LPARs. Facilities are provided for management of the shared resources of the IOA via a Physical Function (PF) of the IOA by assignment of that PF to an I/O Virtualization Management Partition (IMP). The code running in the IMP acts as a virtual intermediary to the VFs for fully managing the VF error handling, VF reset, and configuration operations. The IMP also acts as an interface to the PF for accessing common VF functionality. Furthermore, the functions of resource assignment and management relative to the VFs and the client partitions that use those VFs, which might normally be done by an entity like a hypervisor, are implemented by this IMP. | 06-04-2009 |
20090245242 | Virtual Fibre Channel Over Ethernet Switch - A system for using a virtual switch in a network. In response to a switch connecting to a network, network configuration data is read to determine whether the switch is part of a virtual switch. In response to determining that the switch is part of the virtual switch, a priority number for the switch is sent to other switches connected to the network. Priority numbers are received from the other switches. It is determined whether the switch has a highest priority number. In response to determining that the switch does have the highest priority number, the switch is selected to be a master switch for the virtual switch. Then, domain identification data for the virtual switch is sent to the other switches. Afterward, the switch and the other switches respond to requests over the network as one domain identified by the domain identification data. | 10-01-2009 |
20090249330 | METHOD AND APPARATUS FOR HYPERVISOR SECURITY CODE - Disclosed is a computer implemented method, apparatus, and computer program product for regulating received data in a multiple operating system environment on an I/O adapter. The method includes a hypervisor for determining that the I/O adapter indicated a receive completion. The hypervisor, responsive to retrieving the receive completion, determines that the receive completion is associated with a successful status. The hypervisor, determines in hypervisor space whether an at least one data packet satisfies a security criterion. The hypervisor, routes the data packet to at least one selected from a group consisting of an operating system partition of the multiple operating system environment and a network address on a local area network. | 10-01-2009 |
20090271494 | Method, System and Program Product for Providing Server Security Via A Security Sensor Application Shared by Multiple Operating System Partitions - When a hypervisor in a computer server receives input/output (I/O) data traffic, the hypervisor sends the I/O data traffic to a security sensor application shared by multiple operating system (OS) partitions. If the security sensor application indicates that the I/O data traffic meets pre-defined security standards in the security sensor application, and the I/O data traffic is addressed to one of the OS partitions in the computer server, the hypervisor sends the I/O data traffic to the applicable OS partition. If the I/O data traffic meets the pre-defined security standards, and the I/O data traffic is not addressed to one of the OS partitions, the hypervisor sends the I/O data traffic to an external destination in a network coupled to the computer server. | 10-29-2009 |
20090276526 | Access Control List Endpoint Implementation - A method, system, and computer program product for providing direct communications between FCoE endpoint devices within the same fibre channel network zone. A direct fibre channel (DFC) utility provides an FCoE stack with an exclusive ability to define an Ethertype within an ethertype field of an Ethernet packet with “FCoE”. In addition, the DFC utility enables storage of access control lists (ACLs) containing allowed destination addresses and allowed source addresses within the adapter of an FCoE endpoint. Additionally, the DFC utility initiates an exchange of messages with an Ethernet switch to determine a feasibility of establishing direct connections between endpoints. In particular, the DFC utility determines whether the Ethernet switch supports FCoE ACL checking. Further, the DFC utility creates a zone ID for the FCoE endpoint device. The DFC utility allows direct communication between FCoE endpoints within the same fibre zone. | 11-05-2009 |
20090276544 | Mapping a Virtual Address to PCI Bus Address - Registering memory space within a data processing system is performed. One or more open calls are received from an application to access one or more input/output (I/O) devices. Responsive to receiving the one or more open calls, one or more I/O map and pin calls are sent in order to register memory space for the one or more I/O devices within at least one storage area that will be accessed by the application. At least one virtual I/O bus address is received for each registered memory space of the one or more I/O devices. At least one I/O command is executed using the at least one virtual I/O bus address without intervention by an operating system or operating system image. | 11-05-2009 |
20090276551 | Native and Non-Native I/O Virtualization in a Single Adapter - Mechanisms for enabling both native and non-native input/output virtualization (IOV) in a single I/O adapter are provided. The mechanisms allow a system with a large number of logical partitions (LPARs) and system images to use IOV to share a native IOV enabled I/O adapter or endpoint that does not implement the necessary number of virtual functions (VFs) for each LPAR and system image. A number of VFs supported by the I/O adapter, less one, are assigned to LPARs and system images so that they may make use of native IOV using these VFs. The remaining VF is associated with a virtual intermediary (VI) which handles non-native IOV of the I/O adapter. Any remaining LPARs and system images share the I/O adapter using the non-native IOV via the VI. Thus, any number of LPARs and system images may share the same I/O adapter or endpoint. | 11-05-2009 |
20090276605 | Retaining an Association Between a Virtual Address Based Buffer and a User Space Application that Owns the Buffer - Registering memory space for an application is performed. One or more open calls are received from an application to access one or more input/output (I/O) devices. Responsive to receiving the one or more open calls, one or more I/O map and pin calls are sent in order to register memory space for the one or more I/O devices within at least one storage area that will be accessed by the application. A verification is made as to whether the memory space to be registered is associated with the application. Responsive to the memory space being associated with the application, at least one virtual I/O bus address is received for each registered memory space of the one or more I/O devices. At least one I/O command is executed using the at least one virtual I/O bus address without intervention by an operating system or operating system image. | 11-05-2009 |
20090276773 | Multi-Root I/O Virtualization Using Separate Management Facilities of Multiple Logical Partitions - Mechanisms are provided for implementing a multi-root PCI manager (MR-PCIM) in a multi-root I/O virtualization management partition (MR-IMP) to control the shared functionality of an multi-root I/O virtualization (IOV) enabled switch fabric and multi-root IOV enabled I/O adapter (IOA) through the base functions (BF) of the switches and IOAs. A hypervisor provides device-independent facilities to the code running in the I/O Virtualization Management Partition (IMP), Multi-Root (MR)-IMP and client partitions. The MR-IMP may include device specific code without the hypervisor needing to sacrifice its size, robustness, and upgradeability. The hypervisor provides the virtual intermediary functionally for the sharing and control of the switch and IOA's control functions. | 11-05-2009 |
20090276775 | PCI Function South-Side Data Management - A hypervisor, during device discovery, has code which can examine the south-side management data structure in an adapter's configuration space and determine the type of device which is being configured. The hypervisor may copy the south-side management data structure to a hardware management console (HMC) and the HMC can populate the data structure with south-side data and then pass the structure to the hypervisor to replace the data structure on the adapter. In another embodiment the hypervisor may copy the data structure to the HMC and the HMC can instruct the hypervisor to fill-in the data structure, a virtual function at a time, with south-side management data associations. The administrator can assign south-side data, such as a MAC address for a virtual instance of an Ethernet device, to LPARs sharing the adapter. Thus, a standard way to manage the south-side data of virtual functions is provided. | 11-05-2009 |
20100017497 | Network System with Initiator SubNetwork Communication to Target Subnetwork Communication Including Fibre Channel Over Ethernet to Fibre Channel Over Internet Protocol Conversion - A network system supports multiple network communication protocols. An Ethernet component gateway in a Fibre Channel over Ethernet (FCoE) initiator system converts FCoE data packets from host devices to Fibre Channel over Internet Protocol (FCIP) data packets for transmission to a Storage Area Network (SAN) target system. The SAN target system may include a target Fibre Channel (FC) storage device and a SAN component gateway. The SAN component gateway converts FCIP data packets to SAN data packets for use by the target FC storage device. The SAN data packets may be either FC protocol data packets or FCoE protocol data packets. The SAN target system may provide for discovery of target FC storage device adapter information. | 01-21-2010 |
20100042805 | Method to Perform LUM Masking/Mapping in a SR-IOV Enabled SAS Adapter - A “LUN Table” enables Logical Unit Number (LUN) mapping/masking within an IOV adapter included in a Serial Attached Small Computer System Interface (“SAS” or “Serial Attached SCSI”). A plurality of System Images (“SI”) share block storage through the SAS. The IOV adapter includes one or more Virtual Functions (VF), a Physical Function (PF), and a LUN Table within the PF. The VF allows each SI to communicate I/0 requests with a storage device through the PF. The LUN Table maps the I/0 requests to unique locations within the storage device. Each SI is isolated from all other SIs. Interference between each SI is avoided. A VIOS or a LUN mapping/masking SAN are not required. I/0 latency, processor overhead and storage cost are improved over prior LUN mapping/masking solutions. | 02-18-2010 |
20100097941 | Redundant Intermediary Switch Solution for Detecting and Managing Fibre Channel over Ethernet FCoE Switch Failures - A method for detecting a switch failure and managing switch failover in a Fibre Channel over Ethernet network. The plurality of Fibre Channel over Ethernet switches are connected to an intermediary Ethernet switch and to each other via an Ethernet connection. The intermediary Ethernet switch is different from the plurality of Fibre Channel over Ethernet switches. A plurality of end polls are also connected to the intermediary Ethernet switch. A failure is detected in a first switch in the plurality of Fibre Channel over Ethernet switches. The first switch is associated with a first end port in the plurality of end ports. A failure notification is sent to the first end port using the intermediary Ethernet switch. The failure notification might specify a second switch destination and an identification of the first switch. A fabric login is exchanged between the first end port and the second switch. | 04-22-2010 |
20100103939 | DETERMINING THE CONFIGURATION OF AN ETHERNET FABRIC - A Fibre Channel over Ethernet (FCoE) device obtains knowledge of the configuration of an attached Ethernet fabric during a fabric initialization or login (FLOGI) process. FCoE N_Ports obtain similar information during the port initialization or login (PLOGI) process. The FCoE device may provide network management services to attached FCoE devices. Embodiments include a management module or processor within an FCoE N_Port. The management module or processor receives information associated with an Ethernet fabric from the FCoE N_Ports and provides management services based on the information associated with the Ethernet fabric. The FCoE N_Port communicatively couples to at least one additional FCoE N_Port through the Ethernet fabric. The FCoE N_Ports may be implemented within a disc storage drive, a host bus adapter, and/or an FCoE switch. The capability of an N_Port to obtain this information allows a management module or processor at the N_Port to manage the Ethernet fabric. | 04-29-2010 |
20100104280 | FIBRE CHANNEL FRAMING AND SIGNALING OPTIONAL HEADER FOR ETHERNET FABRIC CONFIGURATION INFORMATION - A Fibre Channel over Ethernet (FCoE) device obtains knowledge of the configuration of an attached Ethernet fabric from data fields within a frame header during a fabric initialization or login (FLOGI) process. FCoE N_Ports obtain similar information during the port initialization or login (PLOGI) process. The FCoE device may provide network management services to attached FCoE devices. Embodiments include a management module or processor within an FCoE N_Port. The management module or processor receives a frame having a header and validates and extracts information associated with an Ethernet fabric from the FCoE N_Ports and in order to provide management services based on the information associated with the Ethernet fabric. The FCoE N_Port communicatively couples to at least one additional FCoE N_Port through the Ethernet fabric. The FCoE N_Ports may be implemented within a disc storage drive, a host bus adapter, and/or an FCoE switch. The capability of an N_Port to obtain this information allows a management module or processor at the N_Port to manage the Ethernet fabric. | 04-29-2010 |
20100146170 | Differentiating Traffic Types in a Multi-Root PCI Express Environment - Mechanisms for differentiating traffic types in a multi-root PCI Express environment are provided. The mechanisms generate a first mapping data structure that, for each single-root virtual hierarchy in the multi-root data processing system, associates a plurality of traffic classes with a plurality of priority groups and maps each traffic class in the plurality of traffic classes to a corresponding virtual channel in a plurality of virtual channels. Moreover, a second mapping data structure is generated that maps each virtual channel in the plurality of virtual channels to corresponding virtual link in a plurality of virtual links of the multi-root data processing system. Traffic of a particular priority group is routed from a single-root virtual hierarchy to a particular virtual link in the plurality of the virtual links based on the first mapping data structure and second mapping data structure. | 06-10-2010 |
20100165874 | Differentiating Blade Destination and Traffic Types in a Multi-Root PCIe Environment - Mechanisms for differentiating traffic types per host system blade in a multi-root PCI Express environment are provided. The mechanisms generate a first mapping data structure that, for each single-root virtual hierarchy in the multi-root data processing system, associates a plurality of traffic classes with a plurality of priority groups and maps each traffic class in the plurality of traffic classes to a corresponding virtual channel in a plurality of virtual channels. Moreover, a second mapping data structure is generated that maps each virtual channel in the plurality of virtual channels to corresponding per host system blade virtual links in a plurality of virtual links of the multi-root data processing system. Traffic of a particular priority group is routed from a single-root virtual hierarchy to a particular virtual link in the plurality of the virtual links based on the first mapping data structure and second mapping data structure. | 07-01-2010 |
20100290467 | Discovery and Capability Exchange Management in a Virtualized Computing Platform Utilizing a SR-IOV Adapter - In various embodiments a traffic class manager is a resource within a virtualized computer systems trusted entity (i.e. a hypervisor, trusted partition, etc.) which maps requirements from a platform management and associated network capabilities onto an SR-IOV adapter in order to appropriately allocate SR-IOV adapter and network resources to the virtualized computer partitions. In various embodiments the traffic class manager defines network traffic classes that meet the objectives of a platform administrator based on the capabilities of the SR-IOV adapter and the network attached to the adapter ports. Once the traffic classes are defined, in various embodiments, the traffic class manager enforces the assignment of a traffic class to a virtual interface queue pair within a partition. | 11-18-2010 |
20120182993 | HYPERVISOR APPLICATION OF SERVICE TAGS IN A VIRTUAL NETWORKING ENVIRONMENT - A physical host executes a virtual machine monitor (VMM) in communication with a plurality of consumer virtual machines (VMs). In response to receipt of a packet, the VMM determines whether a service is to be performed for the packet by a service virtual machine (VM) in communication with the VMM. In response to determining that the service is to be performed for the packet by the service VM, the VMM applies a tag to the packet that differentiates the packet from any other packet sharing a common address with the packet but having a different associated consumer, passes the packet to the service VM for performance of the service, and thereafter removes the tag from the packet in response to receipt of the packet from the service VM following performance of the service. In response to receipt of the packet from the service VM, the VMM forwards the packet. | 07-19-2012 |
20120185846 | NETWORK COMMUNICATIONS OVER SHARED LINKS IN A VIRTUALIZED ENVIRONMENT - A method of data processing includes a physical host executing a virtual machine monitor (VMM) that instantiates a plurality of virtual machines (VMs). The VMM supports processing of a virtual link manager (VLM) that deploys and configures a plurality of Layer 2 virtual links sharing bandwidth of a Layer 2 physical link between the physical host and an access switch. The VMM communicates parameters of the plurality of virtual links with the access switch. | 07-19-2012 |
20120188870 | Discovery and Capability Exchange Management in a Virtualized Computing Platform - In various embodiments a traffic class manager is a resource within a virtualized computer systems trusted entity (e.g. a hypervisor, trusted partition, etc.) that maps requirements from a platform management and associated network capabilities onto an adapter (e.g. SR-IOV adapter, etc.) in order to appropriately allocate adapter and network resources to virtualized computer partitions. In various embodiments the traffic class manager defines network traffic classes that meet the objectives of a platform administrator based on the capabilities of the adapter and the network attached to the adapter ports. Once the traffic classes are defined, in various embodiments, the traffic class manager enforces the assignment of a traffic class to a virtual interface queue pair within a partition. | 07-26-2012 |
20120210319 | NETWORK COMMUNICATIONS OVER SHARED LINKS IN A VIRTUALIZED ENVIRONMENT - A method of data processing includes a physical host executing a virtual machine monitor (VMM) that instantiates a plurality of virtual machines (VMs). The VMM supports processing of a virtual link manager (VLM) that deploys and configures a plurality of Layer 2 virtual links sharing bandwidth of a Layer 2 physical link between the physical host and an access switch. The VMM communicates parameters of the plurality of virtual links with the access switch. | 08-16-2012 |
20120216194 | HYPERVISOR APPLICATION OF SERVICE TAGS IN A VIRTUAL NETWORKING ENVIRONMENT - A physical host executes a virtual machine monitor (VMM) in communication with a plurality of consumer virtual machines (VMs). In response to receipt of a packet, the VMM determines whether a service is to be performed for the packet by a service virtual machine (VM) in communication with the VMM. In response to determining that the service is to be performed for the packet by the service VM, the VMM applies a tag to the packet that differentiates the packet from any other packet sharing a common address with the packet but having a different associated consumer, passes the packet to the service VM for performance of the service, and thereafter removes the tag from the packet in response to receipt of the packet from the service VM following performance of the service. In response to receipt of the packet from the service VM, the VMM forwards the packet. | 08-23-2012 |
20120230192 | LINK LAYER RESERVATION OF SWITCH QUEUE CAPACITY - A network switch, in response to receipt from a source station of a Layer 2 reservation request, establishes a reservation for capacity of an ingress queue of the network switch for a data flow of the source station. In response to a queue overrun condition on the ingress queue of the network switch while the reservation is active, the network switch preserves data frames in the data flow of the source station transmitted pursuant to the reservation and discards other data frames. | 09-13-2012 |
20120230196 | LINK LAYER RESERVATION OF SWITCH QUEUE CAPACITY - A network switch, in response to receipt from a source station of a Layer | 09-13-2012 |
20120230334 | MESSAGE FORWARDING TOWARD A SOURCE END NODE IN A CONVERGED NETWORK ENVIRONMENT - A network node that forwards traffic of a converged network received from a source end node receives a second message addressed to the network node, but intended for the source end node. The second message includes at least a portion of a first message originated by the source end node and previously forwarded by the network node. The network node extracts from the first message a source identifier of the source end node in a first communication protocol and determines by reference to a data structure a destination address of the second message in a second communication protocol. The network node modifies the second message to include the destination address and forwards the second message toward the source end node in accordance with the destination address. | 09-13-2012 |
20120230340 | MESSAGE FORWARDING TOWARD A SOURCE END NODE IN A CONVERGED NETWORK ENVIRONMENT - A network node that forwards traffic of a converged network received from a source end node receives a second message addressed to the network node, but intended for the source end node. The second message includes at least a portion of a first message originated by the source end node and previously forwarded by the network node. The network node extracts from the first message a source identifier of the source end node in a first communication protocol and determines by reference to a data structure a destination address of the second message in a second communication protocol. The network node modifies the second message to include the destination address and forwards the second message toward the source end node in accordance with the destination address. | 09-13-2012 |
20120265840 | Providing a Memory Region or Memory Window Access Notification on a System Area Network - A system and method for providing a memory region/memory window (MR/MW) access notification on a system area network are provided. Whenever a previously allocated MR/MW is accessed, such as via a remote direct memory access (RDMA) read/write operation, a notification of the access is generated and written to a queue data structure associated with the MR/MW. In one illustrative embodiment, this queue data structure may be a MR/MW event queue (EQ) data stricture that is created and used for all consumer processes and all MR/MWs. In other illustrative embodiments, the EQ is associated with a protection domain. In yet another illustrative embodiment, an event record may be posted to an asynchronous event handler in response to the accessing of the MR/MW. In another illustrative embodiment, a previously posted queue element may be used to generate a completion queue element in response to the accessing of the MR/MW. | 10-18-2012 |
20120290695 | Distributed Policy Service - According to one embodiment of the present disclosure, an approach is provided in which a policy server receives a request for a policy from a requestor. The policy server identifies an initiating virtual machine; the initial virtual machine's corresponding virtual network; and a destination virtual machine. Next, a policy corresponding to sending data from the first virtual machine to the second virtual machine is selected. The policy includes one or more logical references to the virtual network and does not include a physical reference to a physical entity located on a physical network. In turn, a physical path translation corresponding to the selected policy is identified and sent to the requestor. | 11-15-2012 |
20120290703 | Distributed Policy Service - According to one embodiment of the present disclosure, an approach is provided in which a policy server receives a request for a policy from a requestor. The policy server identifies an initiating virtual machine; the initial virtual machine's corresponding virtual network; and a destination virtual machine. Next, a policy corresponding to sending data from the first virtual machine to the second virtual machine is selected. The policy includes one or more logical references to the virtual network and does not include a physical reference to a physical entity located on a physical network. In turn, a physical path translation corresponding to the selected policy is identified and sent to the requestor. | 11-15-2012 |
20120291024 | Virtual Managed Network - According to one embodiment of the present disclosure, an approach is provided in which a policy module receives data that is initiated by a first virtual machine and has a destination at a second virtual machine. The policy module selects a policy that corresponds to sending the data from the first virtual machine to the second virtual machine. The policy includes one or more logical references to one or more virtual networks, and does not include a physical reference to a physical entity located on a physical network. In turn, the policy module encapsulates the data with a physical path translation that is based upon the selected policy, and sends the encapsulated data over the physical network to a second policy module that corresponds to the second virtual machine. | 11-15-2012 |
20120297384 | Virtual Managed Network - According to one embodiment of the present disclosure, an approach is provided in which a policy module receives data that is initiated by a first virtual machine and has a destination at a second virtual machine. The policy module selects a policy that corresponds to sending the data from the first virtual machine to the second virtual machine. The policy includes one or more logical references to one or more virtual networks, and does not include a physical reference to a physical entity located on a physical network. In turn, the policy module encapsulates the data with a physical path translation that is based upon the selected policy, and sends the encapsulated data over the physical network to a second policy module that corresponds to the second virtual machine. | 11-22-2012 |
20130033993 | Distributed Overlay Network Data Traffic Management by a Virtual Server - An approach is provided in which a data traffic module executing on a network interface card receives a data packet initiated by a first virtual machine with a destination at a second virtual machine. The data traffic module identifies one or more physical path translations corresponding to a logical connectivity that is independent of physical topology constraints of a physical network. In turn, the data traffic module encapsulates the data packet with the one or more physical path translations and sends the encapsulated data packet to the second virtual machine over the physical network. | 02-07-2013 |
20130034094 | Virtual Switch Data Control In A Distributed Overlay Network - An approach is provided in which a hypervisor provisions switch resources on a network interface card, which includes a virtual switch and a physical port. The hypervisor invokes a switch control module on a virtual machine, which provides control information to one or more of the switch resources. In turn, one or more of the switch resources utilize the control information to direct data packets between a source virtual machine and a destination virtual machine over one or more virtual networks that are independent of physical topology constraints of a physical network. | 02-07-2013 |
20130034109 | Distributed Overlay Network Data Traffic Management by a Virtual Server - An approach is provided in which a data traffic module executing on a network interface card receives a data packet initiated by a first virtual machine with a destination at a second virtual machine. The data traffic module identifies one or more physical path translations corresponding to a logical connectivity that is independent of physical topology constraints of a physical network. In turn, the data traffic module encapsulates the data packet with the one or more physical path translations and sends the encapsulated data packet to the second virtual machine over the physical network. | 02-07-2013 |
20130042086 | Dynamic Network Adapter Memory Resizing and Bounding for Virtual Function Translation Entry Storage - An approach is provided which a system selects a first virtual function from a plurality of virtual functions executing on a network adapter that includes a memory area. Next, the system allocates, in the memory area, a memory corresponding to the first virtual function. The system then stores one or more translation entries in the allocated memory partition, which are utilized to send data traversing through the first virtual function. As such, the system sends, utilizing one or more of the translation entries, the data packets from the network adapter to one or more destinations. In turn, the system dynamically resizes the memory partition based upon an amount of the memory partition that is utilized to store the one or more translation entries. | 02-14-2013 |
20130042237 | Dynamic Network Adapter Memory Resizing and Bounding for Virtual Function Translation Entry Storage - An approach is provided which a system selects a first virtual function from a plurality of virtual functions executing on a network adapter that includes a memory area. Next, the system allocates, in the memory area, a memory corresponding to the first virtual function. The system then stores one or more translation entries in the allocated memory partition, which are utilized to send data traversing through the first virtual function. As such, the system sends, utilizing one or more of the translation entries, the data packets from the network adapter to one or more destinations. In turn, the system dynamically resizes the memory partition based upon an amount of the memory partition that is utilized to store the one or more translation entries. | 02-14-2013 |
20130042238 | Optimized Virtual Function Translation Entry Memory Caching - An approach is provided in which a virtual function, which executes on a network adapter, receives a data packet from a first virtual machine. A translation entry is identified that corresponds to sending the data packet from the first virtual machine to a second virtual machine, and a determination is made as to whether an onboard memory partition assigned to the virtual function includes the identified translation. If the onboard memory location includes the translation entry, the data packet is sent to the destination virtual machine using the translation entry retrieved from the onboard memory partition. Otherwise, if the translation entry is not located in the onboard memory partition, the data packet is sent to the destination virtual machine using a translation entry retrieved from an off board memory location. | 02-14-2013 |
20130042240 | Optimized Virtual Function Translation Entry Memory Caching - An approach is provided in which a virtual function, which executes on a network adapter, receives a data packet from a first virtual machine. A translation entry is identified that corresponds to sending the data packet from the first virtual machine to a second virtual machine, and a determination is made as to whether an onboard memory partition assigned to the virtual function includes the identified translation. If the onboard memory location includes the translation entry, the data packet is sent to the destination virtual machine using the translation entry retrieved from the onboard memory partition. Otherwise, if the translation entry is not located in the onboard memory partition, the data packet is sent to the destination virtual machine using a translation entry retrieved from an off board memory location. | 02-14-2013 |
20130083690 | Network Adapter Hardware State Migration Discovery in a Stateful Environment - An approach is provided in which a discovery system receives a migration request to move a virtual machine that executes on a first system. The discovery system identifies a first network adapter corresponding to the first system, and identifies hardware state data used by the first network adapter to process data packets generated by the virtual machine. In turn, the discovery system identifies a second network adapter that is compatible with a native format of the hardware state data, and migrates the virtual machine to a second system corresponding to the identified second network adapter. | 04-04-2013 |
20130086582 | Network Adapter Hardware State Migration Discovery in a Stateful Environment - An approach is provided in which a discovery system receives a migration request to move a virtual machine that executes on a first system. The discovery system identifies a first network adapter corresponding to the first system, and identifies hardware state data used by the first network adapter to process data packets generated by the virtual machine. In turn, the discovery system identifies a second network adapter that is compatible with a native format of the hardware state data, and migrates the virtual machine to a second system corresponding to the identified second network adapter. | 04-04-2013 |
20130091261 | Defining And Managing Virtual Networks In Multi-Tenant Virtualized Data Centers - An approach is provided in which a computer system selects a virtual domain from multiple virtual domains, which are each overlayed onto a physical network and are independent of physical topology constraints of the physical network. The computer system selects, from the selected virtual domain, a first virtual group that includes one or more first virtual network endpoints. Next, the computer system selects, from the selected virtual domain, a second virtual group that includes one or more second virtual network endpoints. In turn, the computer system creates a logical link policy that includes one or more actions corresponding to sending data between the first virtual group and the second virtual group. | 04-11-2013 |
20130091501 | Defining And Managing Virtual Networks In Multi-Tenant Virtualized Data Centers - An approach is provided in which a computer system selects a virtual domain from multiple virtual domains, which are each overlayed onto a physical network and are independent of physical topology constraints of the physical network. The computer system selects, from the selected virtual domain, a first virtual group that includes one or more first virtual network endpoints. Next, the computer system selects, from the selected virtual domain, a second virtual group that includes one or more second virtual network endpoints. In turn, the computer system creates a logical link policy that includes one or more actions corresponding to sending data between the first virtual group and the second virtual group. | 04-11-2013 |
20130097600 | Global Queue Pair Management in a Point-to-Point Computer Network - An approach is provided in which a local module receives a data frame initiated by a first virtual machine and has a target destination at a second virtual machine, which executes on a destination host system. The local module identifies a destination local port ID and a destination global queue pair number corresponding to the second virtual machine. In one embodiment, the destination local port ID corresponds to the destination host, but the destination global queue pair number is independent of the destination host. The local module includes the destination global queue pair number and the destination local port ID in an overlay header and encapsulates the data frame with the overlay header, which results in an encapsulated frame. In turn, the local module sends the encapsulated frame through a computer network to the second virtual machine. | 04-18-2013 |
20130152075 | Acceleration for Virtual Bridged Hosts - An approach is provided in which a hardware accelerated bridge executing on a network adapter receives an ingress data packet. The data packet includes a destination MAC address that corresponds to a virtual machine, which interfaces to a software bridge executing on a hypervisor. The hardware accelerated bridge identifies a software bridge table entry that includes the destination MAC address and a virtual function identifier, which identifies a virtual function corresponding to the software bridge. In turn, the hardware accelerated bridge sends the data packet from the hardware accelerated bridge to the software bridge through the identified virtual function. | 06-13-2013 |
20130223277 | DISJOINT MULTI-PATHING FOR A DATA CENTER NETWORK - A method for determining disjoint multi-paths in a data center fabric network system, according to one embodiment, includes creating a topology of a data center fabric network system, wherein the system comprises a plurality of switches; determining paths between the plurality of switches; calculating one or more disjoint multi-paths from a source device to a destination device in the system; and writing the one or more disjoint multi-paths into a forwarding table of at least one of the source device and a nearest shared switch to the source device. | 08-29-2013 |
20130223440 | DISJOINT MULTI-PATHING FOR A DATA CENTER NETWORK - According to one embodiment, a data center fabric network system includes a controller, a plurality of switches connected to the controller via a plurality of communication links, a source device connected to at least a first switch of the plurality of switches, a destination device connected to at least a second switch of the plurality of switches, and wherein the controller is adapted for: creating a topology of the data center fabric network system, determining paths between the plurality of switches, calculating one or more disjoint multi-paths from the source device to the destination device based on the topology, and writing the one or more disjoint multi-paths into a forwarding table of at least one of: the source device, the first switch, and a nearest shared switch to the source device. | 08-29-2013 |
20140071983 | OVERLAY NETWORK CAPABLE OF SUPPORTING STORAGE AREA NETWORK (SAN) TRAFFIC - In one embodiment, a system includes a server having a hypervisor layer that includes an overlay-capable virtual switch and a Fibre Channel (FC) virtual switch; at least one processor adapted for executing logic; and a converge network adapter (CNA) coupled to the hypervisor, the CNA being adapted for communicating with the overlay-capable virtual switch and the FC virtual switch, wherein the CNA is adapted for communicating with a FC forwarder (FCF), and the overlay-capable virtual switch includes logic adapted for communicating with a central controller | 03-13-2014 |
20140071990 | OVERLAY NETWORK CAPABLE OF SUPPORTING STORAGE AREA NETWORK (SAN) TRAFFIC - In one embodiment, a method for sending FCoE traffic over an overlay network includes receiving a block of media access control (MAC) addresses for use in local fabric initialization protocol (FIP) operations and advertising to all local FCoE initiators to route FCoE traffic to a server by asserting an overlay-capable virtual switch within the server as a FCF for any requesting FCoE initiators. | 03-13-2014 |
20140096183 | PROVIDING SERVICES TO VIRTUAL OVERLAY NETWORK TRAFFIC - In one embodiment, a method for applying security policy in an overlay network includes receiving a request, including a packet, for a communication path through an overlay network, determining whether a security policy is to be applied to the packet based on at least one of: contents of the packet, first information, and second information, selecting a communication path between a source physical switch and a destination physical switch, wherein the selected communication path directly connects the source physical switch to the destination physical switch when it is determined to not apply the security policy to the packet, and the selected communication path connects the source physical switch to the destination physical switch via a security appliance when it is determined to apply the security policy to the packet, and sending the selected communication path to the source physical switch. | 04-03-2014 |
20140198649 | EXTENDED LINK AGGREGATION (LAG) FOR USE IN MULTIPLE SWITCHES - In one embodiment, a method for providing link aggregation (LAG) to heterogeneous switches includes receiving, at a switch controller, LAG requests forwarded by switches and determining that multiple LAG requests corresponding to a server have been received, grouping the multiple LAG requests into LAG groups according to a switch from which they were received and correlating all the LAG groups with the server, instructing each of the switches to setup a LAG group with the server according to the LAG groups determined by the switch controller, and creating alternate flows that correspond to flows through each of the switches to the server through direction from the switch controller. The switches may rely upon OpenFlow to communicate with the switch controller, in some approaches. In addition, other methods for providing LAG to heterogeneous switches are also described, along with systems and computer program products which provide LAG to heterogeneous switches. | 07-17-2014 |
20140241345 | SOURCE ROUTING WITH FABRIC SWITCHES IN AN ETHERNET FABRIC NETWORK - In one embodiment, a system includes a network fabric having a plurality of fabric switches interconnected in the network fabric and a switch controller having logic adapted to configure the network fabric, determine one or more paths through the network fabric between any two hosts connected thereto, and create a source-routing table to store the one or more paths through the network fabric between any two hosts connected thereto. In another embodiment, a method includes receiving or creating a packet using a NIC of a host connected to a network fabric having a plurality of fabric switches interconnected therein, determining a path through the network fabric by consulting a source-routing table stored to the host, storing source-routing information to a packet header for the packet, the source-routing information including the path, and sending the packet to a first device or hop indicated by the path in the source-routing information. | 08-28-2014 |
20140269274 | CREDIT-BASED FLOW CONTROL IN LOSSLESS ETHERNET NETWORKS - In one embodiment, a system includes a hardware processor and logic integrated with and/or executable by the processor or media access control (MAC) functionality of a network port, the logic being adapted to initialize a link between a receiving endpoint and a sending endpoint, the receiving and sending endpoints being connected in a network fabric, wherein at least one virtual link is created within the link, receive an amount of available flow credits from the receiving endpoint, wherein the amount of available flow credits are used to determine a capacity to process packets at the receiving endpoint, and transmit one or more packets to the receiving endpoint until all packets are sent or the amount of available flow credits is insufficient to process additional packets, wherein exchange of flow credits is performed on a per virtual link basis. | 09-18-2014 |
20140269705 | HETEROGENEOUS OVERLAY NETWORK TRANSLATION FOR DOMAIN UNIFICATION - In one embodiment, a method for providing multi-protocol overlay handling includes receiving first traffic via an input overlay tunnel at a multi-protocol virtual tunnel end point (VTEP)-enabled device, the first traffic including a plurality of overlay-encapsulated packets which adhere to a first overlay network protocol, and wherein the input overlay tunnel adheres to the first overlay network protocol; routing the first traffic to a second overlay network tunnel which adheres to a second overlay network protocol when a destination of the first traffic is specified as the second overlay network tunnel, the second overlay network tunnel being terminated at the multi-protocol VTEP-enabled device; and bridging the first traffic to a destination overlay network tunnel terminated at the multi-protocol VTEP-enabled device when the destination of the first traffic is specified as the destination overlay network tunnel, the destination overlay network tunnel being terminated at the multi-protocol VTEP-enabled device. | 09-18-2014 |
20140301187 | CREDIT-BASED LINK LEVEL FLOW CONTROL AND CREDIT EXCHANGE USING DCBX - In one embodiment, a system includes a network having one or more devices interconnected therebetween, the network including a first device having a first port and logic integrated with and/or executable by a first processor, the logic being adapted to exchange credit exchange (CE) capabilities between the first port and a second port of a second device in the network, receive a credit grant packet from the second port, the credit grant packet indicating credits granted to the first port, and send one or more packets of data from the first port to the second port, the one or more packets corresponding with the granted credits. Other systems and methods for providing credit-based flow control are described according to more embodiments. | 10-09-2014 |
20140348162 | OVERLAY NETWORK CAPABLE OF SUPPORTING STORAGE AREA NETWORK (SAN) TRAFFIC - In one embodiment, a system includes a server having a hypervisor layer that includes an overlay-capable virtual switch and a Fibre Channel (FC) virtual switch; at least one processor configured to execute logic; and a converge network adapter (CNA) coupled to the hypervisor, the CNA being configured to communicate with the overlay-capable virtual switch and the FC virtual switch, wherein the CNA is configured to communicate with a FC forwarder (FCF), and the overlay-capable virtual switch includes logic configured to communicate with a central controller. In another embodiment, a method for sending FCoE traffic over an overlay network includes receiving a block of media access control (MAC) addresses for use in local fabric initialization protocol (FIP) operations and advertising to all local FCoE initiators to route FCoE traffic to a server by asserting an overlay-capable virtual switch within the server as a FCF for any requesting FCoE initiators. | 11-27-2014 |
20150074276 | WORKLOAD DEPLOYMENT WITH REAL-TIME CONSIDERATION OF GLOBAL NETWORK CONGESTION - In one embodiment, a system includes a switch controller configured to communicate with each of a plurality of network devices in a network, the switch controller including a processor and logic integrated with and/or executable by the processor, the logic being configured to: receive a request to instantiate a new workload or to move a current workload in the network, determine a throughput associated with the new or current workload, and select an optimum location to place the new workload or to move the current workload in the network based on the throughput associated with the new or current workload. According to another embodiment, a method includes receiving a request at a switch controller to instantiate or move a workload, determining a throughput associated with the workload, and selecting an optimum location to place or move the workload based on the throughput associated with the workload. | 03-12-2015 |