Patent application number | Description | Published |
20100302940 | LOAD BALANCING ACROSS LAYER-2 DOMAINS - The present application relates to network configurations and specifically to scalable load balancing network configurations. One implementation includes an external client coupled to a scalable load balancing system. The scalable load balancing system includes a load balancing layer that is configured to encapsulate individual incoming packets of a packet flow from the external client. The load balancing layer is further configured to route the incoming packets to target devices on the system. The target devices can span multiple IP subnets. The incoming packets can pass through one or more load balancers of the load balancing layer before reaching individual target devices. Individual target devices can be configured to route at least some outgoing packets of the packet flow to the external client without passing through any of the one or more load balancers. | 12-02-2010 |
20100306408 | AGILE DATA CENTER NETWORK ARCHITECTURE - This patent application relates to an agile network architecture that can be employed in data centers, among others. One implementation provides a virtual layer-2 network connecting machines of a layer-3 infrastructure. | 12-02-2010 |
20110317554 | Distributed and Scalable Network Address Translation - A method of enabling an electronic privately addressable source to be publicly addressable starts at a receiver where an electronic message is received. It is communicated from a sender with a private address outside a subnet of the receiver through a translator. The translator retrieves a lease to at least one of a public address or a port from a lease manager, translates the private address and the private port into a public address and a public port and communicates identifying data such as the public address and the public port to the receiver. If a response is communicated to the private sender, the response may be communicated to the private sender through the network. The public address and the public port on the message may be translated to the private address and the port of the private sender and the private address and the private port may be used to properly route the response to the private sender. | 12-29-2011 |
20110320586 | Flexible and Safe Monitoring of Computers - Described is a server monitoring technology that is scalable to large numbers of servers, e.g., in a datacenter. Agents on servers run queries to monitor data sources for that server, such as performance counters and other events. The agents monitor their resource usage and those of monitored events to stay within an administrator-specified resource budget (policy), e.g., by modifying the running queries and/or monitoring parameters. A controller receives results of the monitoring, analyzes the results, and takes action as needed with respect to server operation and monitoring. The controller may dynamically update an agent's queries, monitoring parameters and/or monitored data post-processing operations. The controller may issue alerts and reports, including alerts indicative of inter-server problems between two or more servers. | 12-29-2011 |
20130142201 | CONNECTING ON-PREMISE NETWORKS WITH PUBLIC CLOUDS - A computer system for encapsulating a packet between a customer premise for delivery to customer resources within a public cloud data center. The computer system comprises a shim gateway. The shim gateway comprises a plurality of customer specific shim components. The shim gateway is configured to receive a packet from a customer premise. The packet has a VLAN tag. The packet identifies a tenant within a designated virtual network for the customer. The designated virtual network is within the public cloud data center. The shim gateway is further configured to encapsulate the packet into an encapsulated packet. Encapsulation includes mapping the VLAN tag to a destination network address of a tenant gateway for the customer. The tenant gateway is in the designated virtual network. The shim gateway is further configured to forward the encapsulated packet to the tenant gateway in the designated virtual network for delivery to the identified tenant. | 06-06-2013 |
Patent application number | Description | Published |
20090059793 | Traffic engineering method, system and computer program product for managing traffic over dynamic networks during both normal and unexpected traffic scenarios - A network traffic engineering method, system and computer program cope with dynamic and unpredictable changes in traffic demands and in the availability and quality of interdomain routes by monitoring traffic over a network having nodes and links, calculating a routing utilizing a convex-hull-based optimal traffic engineering algorithm with penalty envelope (COPE), and adjusting network traffic flow in accordance with the calculated routing. Aggregating collected historical traffic matrices to produce a predicted traffic matrix, the method optimizes for the expected traffic scenario while providing a worst-case guarantee for unexpected traffic scenarios and thereby advantageously achieves efficient resource utilization during normal traffic and avoids network congestion in a wide variety of scenarios. | 03-05-2009 |
20090303880 | DATA CENTER INTERCONNECT AND TRAFFIC ENGINEERING - A system for commoditizing data center networking is disclosed. The system includes an interconnection topology for a data center having a plurality of servers and a plurality of nodes of a network in the data center through which data packets may be routed. The system uses a routing scheme where the routing is oblivious to the traffic pattern between nodes in the network, and wherein the interconnection topology contains a plurality of paths between one or more servers. The multipath routing may be Valiant load balancing. It disaggregates the function of load balancing into a group of regular servers, with the result that load balancing server hardware can be distributed amongst racks in the data center leading to greater agility and less fragmentation. The architecture creates a huge, flexible switching domain, supporting any server/any service, full mesh agility, and unregimented server capacity at low cost. | 12-10-2009 |
20090307334 | DATA CENTER WITHOUT STRUCTURAL BOTTLENECKS - A method of networking a plurality of servers together within a data center is disclosed. The method includes the step of addressing a data packet for delivery to a destination server by providing the destination server address as a flat address. The method further includes the steps of obtaining routing information required to route the packet to the destination server. This routing information may be obtained from a directory service servicing the plurality of servers. Once the routing information is obtained, the data packet may be routed to the destination server according to the flat address of the destination server and routing information obtained from the directory service. | 12-10-2009 |
20090310512 | METHOD AND APPARATUS FOR CREATING ROUTER CONFIGURATION - A method and apparatus for creating one or more router configurations in a network are disclosed. For example, the method receives a request to create a router configuration, and retrieves automatically one or more templates in response to the request. The method then instantiates the router configuration by applying the one or more templates. | 12-17-2009 |
20090310513 | METHOD AND APPARATUS FOR CREATING ROUTER CONFIGURATION TEMPLATES - A method and apparatus for creating router configuration templates are disclosed. For example, the method creates one or more command strings for performing at least one of: a context substitution, or a functional substitution, and stores the one or more command strings as the router configuration template. | 12-17-2009 |
20110211449 | COMMUNICATION TRANSPORT OPTIMIZED FOR DATA CENTER ENVIRONMENT - Methods and apparatus for congestion control in computer networks achieve high burst tolerance, low latency and high throughput with shallow-buffered switches. A method for controlling congestion includes transmitting a set of data packets on a network connection from a first computing device to a second computing device, identifying each data packet in the set of data packets that experienced congestion on the network connection, sending, by the second computing device to the first computing device, a sequence of bits that represents the number of data packets in the set of data packets that were identified as having experienced congestion, and adjusting a rate of transmitting data packets on the network connection based on the sequence of bits sent to the first computing device. | 09-01-2011 |
20110270964 | USING DNS REFLECTION TO MEASURE NETWORK PERFORMANCE - A top level domain name system (DNS) server receives a DNS query from a local DNS resolver, the DNS query requesting a network address corresponding to a domain name. The top level DNS server reflects the local DNS resolver to a reflector DNS server. The reflector DNS server reflects the local DNS resolver to a collector DNS server, which in turn returns the network address to the local DNS resolver. The reflector DNS server and collector DNS server are both in the same data center, and one or more network performance measurements for communications between the local DNS resolver and the data center are determined based on the communications between the local DNS resolver and both the reflector DNS server and the collector DNS server. | 11-03-2011 |
20120127995 | BACKBONE NETWORK WITH POLICY DRIVEN ROUTING - A source device obtains a data packet that includes both a destination address and a payload. The source device selects an exit point address of multiple exit point addresses corresponding to the destination address based on one or more policies. The source device encapsulates the data packet with a header that includes the selected exit point address, and the encapsulated data packet is provided to the backbone network. The encapsulated data packet is routed through the backbone network based on the exit point address, and an edge router of the backbone network identifies an interface of the edge router that corresponds to the exit point address. The header is removed from the encapsulated data packet, and the data packet is added to a buffer of the interface for routing to one or more other devices outside of the backbone network. | 05-24-2012 |
20120166598 | DISTRIBUTED NETWORK COORDINATE SYSTEM - This document relates to a distributed network coordinate system. One implementation provides computer-readable storage media including instructions that may cause a processor to perform certain acts. For example, the acts may include storing an initial network location of a first device in a network. The network may include the first device and a second device. The acts may also include monitoring one or more network performance metrics related to existing application communications with the second device, and determining an updated network location of the first device, based on the initial network location and the network performance metrics. Aspects of network health can be derived from monitoring changes in the network locations of various devices within the coordinate system. | 06-28-2012 |
20130061047 | SECURE AND EFFICIENT OFFLOADING OF NETWORK POLICIES TO NETWORK INTERFACE CARDS - Techniques for efficient and secure implementation of network policies in a network interface controller (NIC) in a host computing device operating a virtualized computing environment. In some embodiments, the NIC may process and forward packets directly to their destinations, bypassing a parent partition of the host computing device. In particular, in some embodiments, the NIC may store network policy information to process and forward packets directly to a virtual machine (VM). If the NIC is unable to process a packet, then the NIC may forward the packet to the parent partition. In some embodiments, the NIC may use an encapsulation protocol to transmit address information in packet headers. In some embodiments, this address information may be communicated by the MC to the parent partition via a secure channel. The NIC may also obtain, and decrypt, encrypted addresses from the VMs for routing packets, bypassing the parent partition. | 03-07-2013 |
20130279899 | METHOD AND APPARATUS FOR PROVIDING A CONTROL PLANE ACROSS MULTIPLE OPTICAL NETWORK DOMAINS - Method and apparatus for providing a control plane across multiple optical network domains is described. In one example, topology and resource discovery is employed in a transport network having a plurality of control domains. Topology information for each of the control domains is summarized to produce summarized topology information. The summarized topology information for each of the control domains is exchanged among the control domains. At each of the control domains, the summarized topology information of each of the control domains is combined to produce combined topology information. The combined topology information is passed among network elements within each of the control domains using an intra-domain routing protocol. | 10-24-2013 |
20130343191 | ENSURING PREDICTABLE AND QUANTIFIABLE NETWORKING PERFORMANCE - The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network. | 12-26-2013 |
20130346465 | APPLICATION ENHANCEMENT USING EDGE DATA CENTER - A management service that receives requests for the cloud computing environment to host applications, and improves performance of the application using an edge server. In response to the original request, the management service allocates the application to run on an origin data center, evaluates the application by evaluating at least one of the application properties designated by an application code author or provider, or the application performance, and uses an edge server to improve performance of the application in response to evaluating the application. For instance, a portion of application code may be offloaded to run on the edge data center, a portion of application data may be cached at the edge data center, or the edge server may add functionality to the application. | 12-26-2013 |
20140337524 | DISTRIBUTED NETWORK COORDINATE SYSTEM - This document relates to a distributed network coordinate system. One implementation provides computer-readable storage media including instructions that may cause a processor to perform certain acts. For example, the acts may include storing an initial network location of a first device in a network. The network may include the first device and a second device. The acts may also include monitoring one or more network performance metrics related to existing application communications with the second device, and determining an updated network location of the first device, based on the initial network location and the network performance metrics. Aspects of network health can be derived from monitoring changes in the network locations of various devices within the coordinate system. | 11-13-2014 |
20140347998 | ENSURING PREDICTABLE AND QUANTIFIABLE NETWORKING PERFORMANCE - The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network. | 11-27-2014 |
Patent application number | Description | Published |
20090089438 | INTELLIGENT NETWORK ADDRESS LOOKUP SERVICE - An intelligent lookup service for a network is provided for clients of a network requesting services of the network that intelligently determines, based on a service requirement of the requested service, optimal service endpoint(s) for providing the requested service. The intelligent lookup service can incorporate predetermined mapping policy and traffic measurements into the determination. In addition, a feedback loop is provided from clients and/or service endpoints to the lookup service concerning measurements about prior connections in the network. The lookup service can include a set of beacons distributed in the network and against which measurements about the network are recorded. A client receives, from the lookup service in response to a request for a network address, a set of candidate service endpoints that pertain to the requested network address and the client connects to one of the candidate service endpoints based on policy or context. | 04-02-2009 |
20100036903 | DISTRIBUTED LOAD BALANCER - Systems and methods that distribute load balancing functionalities in a data center. A network of demultiplexers and load balancer servers enable a calculated scaling and growth operation, wherein capacity of load balancing operation can be adjusted by changing the number of load balancer servers. Accordingly, load balancing functionality/design can be disaggregated to increase resilience and flexibility for both the load balancing and switching mechanisms of the data center. | 02-11-2010 |
20100082804 | MEASURED CLIENT EXPERIENCE FOR COMPUTER NETWORK - Routing network traffic on a computer network is described. In one embodiment, a method is presented which includes transmitting instructions to a client, the instructions executable by the client to request content from two or more content servers, measure two or more network performance characteristics associated with the two or more different content servers, and issue a report to an administrative server. The report may include a first network performance characteristic and a second network performance characteristic. The method may further include selecting a target content server from one of two or more content servers based on comparison of the two or more network performance characteristics; and transmitting routing instructions to an intermediate routing system, the routing instructions executable by the intermediate routing system to direct subsequent content requests transmitted by the client to the target content server. | 04-01-2010 |
20100299349 | Reducing Latency in Returning Online Search Results - An embodiment of the invention is directed to reducing search-response latency. The closest intermediate server can be located between a client computing device and a search engine. A search query is sent to the intermediate server in a first packet of a transport protocol handshake. A plurality of packets are received from the intermediate server. The plurality of packets are used to open a window associated with a transport protocol. A response related to the search query is received by the client. | 11-25-2010 |
20120155468 | MULTI-PATH COMMUNICATIONS IN A DATA CENTER ENVIRONMENT - Various technologies related to multi-path communications in a data center environment are described herein. Network infrastructure devices communicate traffic flows amongst one another, wherein a traffic flow includes a plurality of data packets intended for a particular recipient computing device that are desirably transmitted and received in a certain sequence. Indications that data packets in the traffic flow have been received outside of the certain sequence are processed in a manner to prevent a network infrastructure device from retransmitting a particular data packet. | 06-21-2012 |
20130254766 | OFFLOADING PACKET PROCESSING FOR NETWORKING DEVICE VIRTUALIZATION - The present invention extends to methods, systems, and computer program products for offloading packet processing for networking device virtualization. A host maintains rule set(s) for a virtual machine, and a physical network interface card (NIC) maintains flow table(s) for the virtual machine. The physical NIC receives and processes a network packet associated with the virtual machine. Processing the network packet includes the physical NIC comparing the network packet with the flow table(s) at the physical NIC. When the network packet matches with a flow in the flow table(s) at the physical NIC, the physical NIC performs an action on the network packet based on the matching flow. Alternatively, when the network packet does not match with a flow in the flow table(s) at the physical NIC, the physical NIC passes the network packet to the host partition for processing against the rule set(s). | 09-26-2013 |
20140082048 | NETWORK SERVICES PROVIDED IN CLOUD COMPUTING ENVIRONMENT - A cloud computing environment providing a network service for a client computing entity. The network service is not an application level service, but rather a service that operates at or below the network layer in the protocol stack. For instance, the network service might be a network endpoint service such as a network address service (such as DNS) or a dynamic network service (such as DHCP), or a network traffic service such as a firewall service or a secure tunneling service (such as VPN). The service might also provide a pipeline of network services for network level traffic to and from the client computing entity. The cloud environment uses policy to determine which of a plurality of communication channels to use when exchanging cloud service data for the network service. | 03-20-2014 |
Patent application number | Description | Published |
20090327328 | Platform-Independent Data Application Description Language - Technologies are described herein for generating a platform-independent data application. In one method, a definition of a data schema is received for defining data types. A definition of a user interface is also received for access data stored in the data types. Further, a definition of business logic is received for enforcing validation rules and actions associated with the data types. The definitions of the data schema, the user interface, and the business logic are each received in a platform-independent format and used to generate the platform-independent data application. | 12-31-2009 |
20100199185 | COMMON NAVIGATION MECHANISM FOR DESKTOP AND BROWSER-BASED APPLICATIONS - A client application provides a user interface for defining a form that includes a navigation control. The navigation control includes two or more selectable tabs that are each associated with a target, such as another form or a report. When one of the selectable tabs on the navigation control is selected using an appropriate user input device, the target associated with the selected tab is displayed in the form. The client application also provides functionality for publishing a form including navigation controls to a server application, thereby making the form and the navigation control accessible through a Web interface. | 08-05-2010 |
20100299347 | DATABASE APPLICATION NAVIGATION - A database server application program is provided that is configured to provide a programmable interface into a database application by way of database services uniform resource locators (URLs). A database services URL utilized by the database application can be updated programmatically by program code executing within or under control of the database server application program. A macro action for use in conjunction with a database server application that provides functionality for displaying a database object, such as a form or report, in-place in a Web browser is also disclosed. | 11-25-2010 |
20120299926 | ADAPTIVE TIMELINE VIEWS OF DATA - Concepts and technologies are described herein for providing adaptive timeline views of data. In accordance with the concepts and technologies disclosed herein, data is displayed in an adaptive timeline view in which data is presented using a time scale. The time scale can be modified to increase or decrease a span of time displayed by the adaptive timeline view, or a span of time can be shifted without affecting the time scale applied to the data. Motion can be used to convey information about the data during transitions from a first view to a modified view. As such, the adaptive timeline view can convey various layers of information to a user interacting with the adaptive timeline view. | 11-29-2012 |
20130073589 | DYNAMIC QUERY FOR EXTERNAL DATA CONNECTIONS - An application module generates a query that may involve retrieving data from a local data storage source or from an external data storage source. The application module retrieves metadata to form a query tree stored in the local data storage source and uses the query tree to ascertain one or more sub-queries necessary to complete the query and whether the necessary data for each sub-query is stored locally or externally. Based on the location of the data to complete a sub-query, the application module uses the services of an internal query processing layer if external data is involved for the sub-query. The application module uses the services of a query processing layer in the local data storage source if the data is local. Once all the sub-queries associated with the query tree are performed, then the overall query results are returned to the application module. | 03-21-2013 |