Patent application number | Description | Published |
20080208526 | Strategies for Identifying Anomalies in Time-Series Data - A strategy is described for identifying anomalies in time-series data. The strategy involves dividing the time-series data into a plurality of collected data segments and then using a modeling technique to fit local models to the collected data segments, Large deviations of the time-series data from the local models are indicative of anomalies In one approach, the modeling technique can use an absolute value (L | 08-28-2008 |
20080209030 | Mining Web Logs to Debug Wide-Area Connectivity Problems - Internet service providers and their clients communicate by transmitting messages across one or more networks and infrastructure components. At various points between the service provider and the clients, inclusively, records may be created of each messages occurrence and status. These records may be read and analyzed to determine the effects of the networks and infrastructure components on the provided quality of service. User-effecting incidents (e.g., failures) occurring at networks may also be identified and described. | 08-28-2008 |
20080209273 | Detect User-Perceived Faults Using Packet Traces in Enterprise Networks - Exemplary methods, computer-readable media, and systems for detecting a fault by a packet trace, includes monitoring at least one packet transmitted to or received from, an computing device of an end user, between one or more computing devices implementing at least one of a service or an application on an enterprise network. The process also includes identifying whether an abnormal condition occurred on the computing device of the end user based on monitoring at least one packet transmitted to or received from, the computing device of the end user; and detecting a fault by using an algorithm based on monitoring at least one packet transmitted or received from, the computing device of the end user; wherein the fault indicates a desired course of action did not occur while the computing device of the end user uses at least one of the service or the application in the enterprise network. | 08-28-2008 |
20080222068 | Inferring Candidates that are Potentially Responsible for User-Perceptible Network Problems - Candidates that are potentially responsible for user-perceptible network problems may be inferred. In an example embodiment, a system includes an inference engine to produce a list of candidates that are potentially responsible for user-perceptible network problems, with the candidates being network components that may include both services and network elements. A response to a service request may be a non response, an incorrect response, an untimely correct response, or a timely correct response. The user-perceptible network problems may include the untimely correct response as well as the non response and the incorrect response. In another example embodiment, a method includes monitoring a network and producing a list of candidates that are potentially responsible for user-perceptible network problems. The candidates of the list may include both services and network elements of the monitored network. | 09-11-2008 |
20080222287 | Constructing an Inference Graph for a Network - Constructing an inference graph relates to the creation of a graph that reflects dependencies within a network. In an example embodiment, a method includes determining dependencies among components of a network and constructing an inference graph for the network responsive to the dependencies. The components of the network include services and hardware components, and the inference graph reflects cross-layer components including the services and the hardware components. In another example embodiment, a system includes a service dependency analyzer and an inference graph constructor. The service dependency analyzer is to determine dependencies among components of a network, the components including services and hardware components. The inference graph constructor is to construct an inference graph for the network responsive to the dependencies, the inference graph reflecting cross-layer components including the services and the hardware components. | 09-11-2008 |
20090303880 | DATA CENTER INTERCONNECT AND TRAFFIC ENGINEERING - A system for commoditizing data center networking is disclosed. The system includes an interconnection topology for a data center having a plurality of servers and a plurality of nodes of a network in the data center through which data packets may be routed. The system uses a routing scheme where the routing is oblivious to the traffic pattern between nodes in the network, and wherein the interconnection topology contains a plurality of paths between one or more servers. The multipath routing may be Valiant load balancing. It disaggregates the function of load balancing into a group of regular servers, with the result that load balancing server hardware can be distributed amongst racks in the data center leading to greater agility and less fragmentation. The architecture creates a huge, flexible switching domain, supporting any server/any service, full mesh agility, and unregimented server capacity at low cost. | 12-10-2009 |
20090307334 | DATA CENTER WITHOUT STRUCTURAL BOTTLENECKS - A method of networking a plurality of servers together within a data center is disclosed. The method includes the step of addressing a data packet for delivery to a destination server by providing the destination server address as a flat address. The method further includes the steps of obtaining routing information required to route the packet to the destination server. This routing information may be obtained from a directory service servicing the plurality of servers. Once the routing information is obtained, the data packet may be routed to the destination server according to the flat address of the destination server and routing information obtained from the directory service. | 12-10-2009 |
20100080144 | RESILIENT 1:N FIRST-HOP GATEWAY SELECTION MECHANISM - Exemplary methods, systems, and computer program products describe selecting a gateway based on health and performance information of a plurality of gateways. The techniques describe gateways advertising health and performance information, computing devices creating a table of this health and performance information, and selecting a gateway using the table. In response to changes in the health and performance information, the computing device may select a different gateway. The process allows network traffic load to be distributed across a plurality of gateways. This process further provides resilience by allowing a plurality of active gateways to substitute for a non-functioning gateway. | 04-01-2010 |
20100299349 | Reducing Latency in Returning Online Search Results - An embodiment of the invention is directed to reducing search-response latency. The closest intermediate server can be located between a client computing device and a search engine. A search query is sent to the intermediate server in a first packet of a transport protocol handshake. A plurality of packets are received from the intermediate server. The plurality of packets are used to open a window associated with a transport protocol. A response related to the search query is received by the client. | 11-25-2010 |
20100318609 | BRIDGING ENTERPRISE NETWORKS INTO CLOUD - An enterprise namespace may be extended into a cloud of networked resources. A portion of the cloud may be dynamically partitioned, and the extension of the enterprise namespace established within the portion. Cloud resources thus remain as easily accessible to enterprise users as those which are physically located on the enterprise network. Thus, components such as applications, virtual machine instantiations, application states, server states, etc., may be easily migrated between the enterprise network and the cloud. | 12-16-2010 |
20110211449 | COMMUNICATION TRANSPORT OPTIMIZED FOR DATA CENTER ENVIRONMENT - Methods and apparatus for congestion control in computer networks achieve high burst tolerance, low latency and high throughput with shallow-buffered switches. A method for controlling congestion includes transmitting a set of data packets on a network connection from a first computing device to a second computing device, identifying each data packet in the set of data packets that experienced congestion on the network connection, sending, by the second computing device to the first computing device, a sequence of bits that represents the number of data packets in the set of data packets that were identified as having experienced congestion, and adjusting a rate of transmitting data packets on the network connection based on the sequence of bits sent to the first computing device. | 09-01-2011 |
20110258290 | Bandwidth-Proportioned Datacenters - A system including at least one storage node and at least one computation node connected by a switch is described herein. Each storage node has one or more storage units and one or more network interface components, the collective bandwidths of the storage units and the network interface components being proportioned to one another to enable communication to and from other nodes at the collective bandwidth of the storage units. Each computation node has logic configured to make requests of storage nodes, an input/output bus, and one or more network interface components, the bandwidth of the bus and the collective bandwidths of the network interface components being proportioned to one another to enable communication to and from other nodes at the bandwidth of the input/output bus. | 10-20-2011 |
20110317554 | Distributed and Scalable Network Address Translation - A method of enabling an electronic privately addressable source to be publicly addressable starts at a receiver where an electronic message is received. It is communicated from a sender with a private address outside a subnet of the receiver through a translator. The translator retrieves a lease to at least one of a public address or a port from a lease manager, translates the private address and the private port into a public address and a public port and communicates identifying data such as the public address and the public port to the receiver. If a response is communicated to the private sender, the response may be communicated to the private sender through the network. The public address and the public port on the message may be translated to the private address and the port of the private sender and the private address and the private port may be used to properly route the response to the private sender. | 12-29-2011 |
20110320586 | Flexible and Safe Monitoring of Computers - Described is a server monitoring technology that is scalable to large numbers of servers, e.g., in a datacenter. Agents on servers run queries to monitor data sources for that server, such as performance counters and other events. The agents monitor their resource usage and those of monitored events to stay within an administrator-specified resource budget (policy), e.g., by modifying the running queries and/or monitoring parameters. A controller receives results of the monitoring, analyzes the results, and takes action as needed with respect to server operation and monitoring. The controller may dynamically update an agent's queries, monitoring parameters and/or monitored data post-processing operations. The controller may issue alerts and reports, including alerts indicative of inter-server problems between two or more servers. | 12-29-2011 |
20120127995 | BACKBONE NETWORK WITH POLICY DRIVEN ROUTING - A source device obtains a data packet that includes both a destination address and a payload. The source device selects an exit point address of multiple exit point addresses corresponding to the destination address based on one or more policies. The source device encapsulates the data packet with a header that includes the selected exit point address, and the encapsulated data packet is provided to the backbone network. The encapsulated data packet is routed through the backbone network based on the exit point address, and an edge router of the backbone network identifies an interface of the edge router that corresponds to the exit point address. The header is removed from the encapsulated data packet, and the data packet is added to a buffer of the interface for routing to one or more other devices outside of the backbone network. | 05-24-2012 |
20120155468 | MULTI-PATH COMMUNICATIONS IN A DATA CENTER ENVIRONMENT - Various technologies related to multi-path communications in a data center environment are described herein. Network infrastructure devices communicate traffic flows amongst one another, wherein a traffic flow includes a plurality of data packets intended for a particular recipient computing device that are desirably transmitted and received in a certain sequence. Indications that data packets in the traffic flow have been received outside of the certain sequence are processed in a manner to prevent a network infrastructure device from retransmitting a particular data packet. | 06-21-2012 |
20120260336 | NETWORK ACCOUNTABILITY AMONG AUTONOMOUS SYSTEMS - In one kind of DoS attack, malicious customers may try to send a large number of filter requests against an innocent customer. In one implementation, a Filter Request Server (FRS) may allow a customer against who a filter request is made to dispute the implicit accusation of the filter request or stop sending malicious traffic. If the customer claims innocence, the FRS may log destination addresses of data packets sent by the customer and identify and ignore false filter requests if these filter requests come from customers who do not correspond to one or more of the destination addresses that have previously been logged by the FRS. | 10-11-2012 |
20130047151 | VIRTUALIZATION GATEWAY BETWEEN VIRTUALIZED AND NON-VIRTUALIZED NETWORKS - Methods and apparatus are provided for controlling communication between a virtualized network and non-virtualized entities using a virtualization gateway. A packet is sent by a virtual machine in the virtualized network to a non-virtualized entity. The packet is routed by the host of the virtual machine to a provider address of the virtualization gateway. The gateway translates the provider address of the gateway to a destination address of the non-virtualized entity and sends the packet to the non-virtualized entity. The non-virtualized entity may be a physical resource, such as a physical server or a storage device. The physical resource may be dedicated to one customer or may be shared among customers. | 02-21-2013 |
20130054782 | DETERMINATION OF UNAUTHORIZED CONTENT SOURCES - A plurality of network addresses from a distributed client is obtained, at least a first portion of the obtained network addresses including resolved network address responses to distributed client requests for resolved network addresses corresponding to one or more network location indicators associated with a first web service. Test content is obtained, based on one or more of the network addresses included in the first portion. It is determined whether the obtained test content includes unauthorized content. | 02-28-2013 |
20130343399 | OFFLOADING VIRTUAL MACHINE FLOWS TO PHYSICAL QUEUES - The present invention extends to methods, systems, and computer program products for offloading virtual machine flows to physical queues. A computer system executes one or more virtual machines, and programs a physical network device with one or more rules that manage network traffic for the virtual machines. The computer system also programs the network device to manage network traffic using the rules. In particular, the network device is programmed to determine availability of one or more physical queues at the network device that are usable for processing network flows for the virtual machines. The network device is also programmed to identify network flows for the virtual machines, including identifying characteristics of each network flow. The network device is also programmed to, based on the characteristics of the network flows and based on the rules, assign one or more of the network flows to at least one of the physical queues. | 12-26-2013 |
20130346465 | APPLICATION ENHANCEMENT USING EDGE DATA CENTER - A management service that receives requests for the cloud computing environment to host applications, and improves performance of the application using an edge server. In response to the original request, the management service allocates the application to run on an origin data center, evaluates the application by evaluating at least one of the application properties designated by an application code author or provider, or the application performance, and uses an edge server to improve performance of the application in response to evaluating the application. For instance, a portion of application code may be offloaded to run on the edge data center, a portion of application data may be cached at the edge data center, or the edge server may add functionality to the application. | 12-26-2013 |
20130346558 | DELIVERY CONTROLLER BETWEEN CLOUD AND ENTERPRISE - A delivery controller for use in an enterprise environment that communicates with a cloud computing environment that is providing a service for the enterprise. As the cloud service processing progresses, some cloud service data is transferred from the cloud computing environment to the enterprise environment, and vice versa. The cloud service data may be exchanged over any one of a number of different types of communication channels. The delivery controller selects which communication channel to use to transfer specific data, depending on enterprise policy. Such policy might consider any business goals of the enterprise, and may be applied at the application level. | 12-26-2013 |
20130346968 | Automated controlling of host over network - The provisioning of a host computing system by a controller located over a wide area network. The host computing system has power-on code that automatically executes upon powering up, and causes the host to notify the controller of the host address. In a first level of bootstrapping, the controller instructs the host to download a maintenance operating system. The host responds by downloading and installing a maintenance operating system, enabling further bootstrapping. The persistent memory may further have security data, such as a public key, that allows the host computing system to securely identify the source of the download instructions (and subsequent instructions) as originating from the controller. A second level of bootstrapping may accomplish the configuring of the host with a hypervisor and a host agent. A third level of bootstrapping may accomplish the provisioning of virtual machines on the host. | 12-26-2013 |
20140078882 | Automated Datacenter Network Failure Mitigation - The subject disclosure is directed towards a technology that automatically mitigates datacenter failures, instead of relying on human intervention to diagnose and repair the network. Via a mitigation pipeline, when a network failure is detected, a candidate set of components that are likely to be the cause of the failure is identified, with mitigation actions iteratively targeting each component to attempt to alleviate the problem. The impact to the network is estimated to ensure that the redundancy present in the network will be able to handle the mitigation action without adverse disruption to the network. | 03-20-2014 |
20140082048 | NETWORK SERVICES PROVIDED IN CLOUD COMPUTING ENVIRONMENT - A cloud computing environment providing a network service for a client computing entity. The network service is not an application level service, but rather a service that operates at or below the network layer in the protocol stack. For instance, the network service might be a network endpoint service such as a network address service (such as DNS) or a dynamic network service (such as DHCP), or a network traffic service such as a firewall service or a secure tunneling service (such as VPN). The service might also provide a pipeline of network services for network level traffic to and from the client computing entity. The cloud environment uses policy to determine which of a plurality of communication channels to use when exchanging cloud service data for the network service. | 03-20-2014 |
20140173157 | COMPUTING ENCLOSURE BACKPLANE WITH FLEXIBLE NETWORK SUPPORT - Computing unit enclosures are often configured to connect units (e.g., server racks or trays) with a wired network. Because the network type may vary (e.g., Ethernet, InfiniBand, and Fibre Channel), such enclosures often provide network resources connecting each unit with each supported network type. However, such architectures may present inefficiencies such as unused network resources, and may constrain network support for the units to a small set of supported network types. Presented herein are enclosure architectures enabling flexible and efficient network support by including a backplane comprising a backplane bus that exchanges data between the units and a network adapter using an expansion bus protocol, such as PCI-Express. By shifting the point of network specialization from the enclosure to the network adapter, such architectures may be compatible with network adapters of any network type that communicate with the units according to a widely supported and network-type-independent expansion bus protocol. | 06-19-2014 |
20140280669 | Memory Sharing Over A Network - Memory is shared among physically distinct, networked computing devices. Each computing device comprises a Remote Memory Interface (RMI) accepting commands from locally executing processes and translating such commands into forms transmittable to a remote computing device. The RMI also accepts remote communications directed to it and translates those into commands directed to local memory. The amount of storage capacity shared is informed by a centralized controller, either a single controller, a hierarchical collection of controllers, or a peer-to-peer negotiation. Requests that are directed to remote high-speed non-volatile storage media are detected or flagged and the process generating the request is suspended such that it can be efficiently revived. The storage capacity provided by remote memory is mapped into the process space of processes executing locally. | 09-18-2014 |
20140280708 | Network Routing Modifications For Distribution Of Data - The transmission of multiple copies of data to other computing devices is optimized by minimizing the number of copies of such data transmitted through an expensive portion of the network. A store-and-forward methodology is utilized to transmit only a single copy through the expensive portion and the data is subsequently forked into multiple copies directed to multiple destination computing devices. Computing devices that are not intended destinations can be conscripted as intermediate computing devices, if appropriate to minimize copies of the data transmitted through an expensive portion. Additionally, accommodation can be made for data that is intolerant of out-of-order delivery by utilizing adaptive protocols that avoid mechanisms that may result in out-of-order delivery for data intolerant of such and by utilizing packet sorting at data convergence points to reorder the data. Different protocol settings can be utilized to transmit data across different portions of the network. | 09-18-2014 |
20140281019 | Network Transmission Adjustment Based On Application-Provided Transmission Metadata - Application-provided transmission metadata is utilized, in conjunction with current network information, to adjust network transmissions. An interface between applications seeking to transmit data and networking components enables the application to provide destination information, communication type information, information regarding the quantity of data to be transferred, timeliness information, data location information, cost information, and other like transmission metadata. Current network information can be obtained by the networking components themselves, or can be provided by, or enhanced by, a centralized controller. The networking components can then optimize both the routing and the protocol settings in the form of adjustments to error control settings, flow control settings, receiver control settings, segmentation settings, and other like protocol settings. | 09-18-2014 |