Patent application number | Description | Published |
20080285435 | INTELLIGENT FAILBACK IN A LOAD-BALANCED NETWORKING ENVIRONMENT - One embodiment of the present invention sets forth a method for failing back network connections to a network interface card (NIC) within a computing device. The method includes the steps of monitoring a failed or unreliable NIC within the computing device, determining that the failed or unreliable NIC has recovered, determining that a functional NIC within the computing device is overloaded, selecting a first connection set communicating through the overloaded NIC, and transferring the first connection set to the recovered NIC. With this approach, intelligent decisions can be advantageously made regarding whether to fail back a network connection set to a recovered NIC based on the traffic loads on the overloaded NIC and the recovered NIC. Such an approach to balancing network traffic across the functional NICs within a computing device may substantially improve overall performance relative to prior art techniques. | 11-20-2008 |
20080285441 | INTELLIGENT LOAD BALANCING AND FAILOVER OF NETWORK TRAFFIC - A hash engine in a network device driver maintains data on the utilization and error rate for each network interface card (“NIC”) within a local computing device. From this data, the hash engine intelligently selects transmit NICs and receive NICs based on various networking parameters provided from a software driver program. Transmit packets sent from the operating system in a local computing device to a remote computing device are intercepted, modified and redirected to transmit NICs selected by the hash engine for transmission to remote computing devices. Similarly, address resolution protocol (“ARP”) response packets sent by the operating system in response to ARP request packets are intercepted, modified and redirected to receive NICs selected by the hash engine for transmission. By selecting receive NICs and transmit NICs in this fashion, the hash engine is able to intelligently load balance transmit and receive traffic in the local computing device, thereby improving overall network performance relative to prior art techniques. | 11-20-2008 |
20080285448 | INTELLIGENT LOAD BALANCING AND FAILOVER OF NETWORK TRAFFIC - A hash engine in a network device driver maintains data on the utilization and error rate for each network interface card (“NIC”) within a local computing device. From this data, the hash engine intelligently selects transmit NICs and receive NICs based on various networking parameters provided from a software driver program. Transmit packets sent from the operating system in a local computing device to a remote computing device are intercepted, modified and redirected to transmit NICs selected by the hash engine for transmission to remote computing devices. Similarly, address resolution protocol (“ARP”) response packets sent by the operating system in response to ARP request packets are intercepted, modified and redirected to receive NICs selected by the hash engine for transmission. By selecting receive NICs and transmit NICs in this fashion, the hash engine is able to intelligently load balance transmit and receive traffic in the local computing device, thereby improving overall network performance relative to prior art techniques. | 11-20-2008 |
20080285472 | INTELLIGENT FAILOVER IN A LOAD-BALANCED NETWORK ENVIRONMENT - A hash table in the network device driver maintains data on the traffic characteristics for each network interface (“NIC”) within a computing device. If one of the NICs in the computing device becomes unreliable, the cost function in the hash engine allows the software driver to initiate network traffic redistribution among the remaining reliable NICs in the computing device. Using this hash engine, the software driver is able to intelligently redirect each of the network connections on an unreliable NIC to a reliable NIC within the computing device, in a way that optimizes the distribution of network traffic across the remaining reliable NICs. Alternatively, if a connection is moved from an old NIC to a new NIC, the software driver can detect the moved connection and offload the moved connection to a hardware offload engine on the new NIC. With this approach, issues such as network interface overloading and computing device performance degradation may be more easily avoided when failing over network connections, thereby improving overall system performance relative to prior art techniques. | 11-20-2008 |
20080285552 | INTELLIGENT FAILOVER IN A LOAD-BALANCED NETWORKING ENVIRONMENT - A hash table in the network device driver maintains data on the traffic characteristics for each network interface (“NIC”) within a computing device. If one of the NICs in the computing device becomes unreliable, the cost function in the hash engine allows the software driver to initiate network traffic redistribution among the remaining reliable NICs in the computing device. Using this hash engine, the software driver is able to intelligently redirect each of the network connections on an unreliable NIC to a reliable NIC within the computing device, in a way that optimizes the distribution of network traffic across the remaining reliable NICs. Alternatively, if a connection is moved from an old NIC to a new NIC, the software driver can detect the moved connection and offload the moved connection to a hardware offload engine on the new NIC. With this approach, issues such as network interface overloading and computing device performance degradation may be more easily avoided when failing over network connections, thereby improving overall system performance relative to prior art techniques. | 11-20-2008 |
20080285553 | INTELLIGENT LOAD BALANCING AND FAILOVER OF NETWORK TRAFFIC - A hash engine in a network device driver maintains data on the utilization and error rate for each network interface card (“NIC”) within a local computing device. From this data, the hash engine intelligently selects transmit NICs and receive NICs based on various networking parameters provided from a software driver program. Transmit packets sent from the operating system in a local computing device to a remote computing device are intercepted, modified and redirected to transmit NICs selected by the hash engine for transmission to remote computing devices. Similarly, address resolution protocol (“ARP”) response packets sent by the operating system in response to ARP request packets are intercepted, modified and redirected to receive NICs selected by the hash engine for transmission. By selecting receive NICs and transmit NICs in this fashion, the hash engine is able to intelligently load balance transmit and receive traffic in the local computing device, thereby improving overall network performance relative to prior art techniques. | 11-20-2008 |
20120155249 | TECHNIQUE FOR IDENTIFYING A FAILED NETWORK INTERFACE CARD WITHIN A TEAM OF NETWORK INTERFACE CARDS - A method for identifying a failed network interface card in a system having two NICs configured as a team includes the steps of transmitting a first data packet from the first NIC to a third NIC, wherein the third NIC is not a member of the team, and transmitting a second data packet from the first NIC to the second NIC or from the second NIC to the third NIC, depending on whether the third NIC responds to the transmission of the first data packet. One advantage of the disclosed method is that it specifically identifies which NIC within the team has failed, which is something that cannot be determined by simply exchanging packets between the two NICs. | 06-21-2012 |
20130121221 | Reducing Power Consumption In Wireless Network Stations By Optimizing Contention Period Overhead With Station Grouping, Proxy CSMA, And TIM Monitoring - A method of saving power in a wireless network can include determining a plurality of stations associated with an AP. The AP can create station groups using group selection logic. Notably, the group selection logic is transparent to the plurality of stations. A plurality of TIMs can then be sent, each TIM allowing only one station group access to a channel during a predetermined time interval, such as a beacon interval. In another method, a station can determine its sleep duration based on at least one of first information from the TIM to generate random sleep duration, second information regarding previous operation of the station, and third information regarding a status of the station. The first, second, and third information can include the number of stations associated with the AP and having buffered data based on the TIM, historical collisions, and power status. | 05-16-2013 |
20140029505 | SYSTEM AND METHOD FOR LOCAL FLOW CONTROL IN A COMMUNICATIONS DEVICE - This specification is directed to systems and methods including a communications device for relaying data packets to at least one receiving node, having a host processor for receiving data packets and a target interface for transmitting the packets over a communications medium, wherein the target interface is configured to extract address and traffic class information from a packet and return the packet to the host processor when it is determined a node is unavailable and wherein the host processor is configured buffer the packet in a queue corresponding to the address information and the traffic class information. | 01-30-2014 |
20140037091 | SYSTEM AND METHOD FOR HYBRID MULTIPLE SOURCE DECRYPTION - Systems and methods are disclosed to improve performance of a communications device receiving encrypted communications from multiple sources. The communications device is configured to increase the amount of decryption performed using a hardware-based process as compared to a software-based process by reprogramming the hardware to store a shared security key corresponding to a frame received from a source, allowing the hardware-based process to decrypt subsequent frames from that source. | 02-06-2014 |
Patent application number | Description | Published |
20110242968 | Method and Apparatus for Improving Network Communication Using BFD and VRRP Tracking System - An apparatus and method of a network system between a host and a group of routers using virtual router redundancy protocol (“VRRP”) messages and bidirectional forwarding detection (“BFD”) sessions are disclosed. The network system is capable of facilitating a first communication between a host and a master router of multiple VRRP routers and establishing a BFD session between the host and the master router. When the BFD session fails, the priority of the master router is subsequently lowered and a backup router is activated. In one embodiment, the backup router capable of performing functions of the master router becomes a new master until the BFD session resumes. | 10-06-2011 |
20140241147 | SYSTEM AND METHOD FOR RAPID LINK FAILURE HANDLING - A system and method for link failure handling includes detecting a failure in a first network connection between a first network switching unit and a second network switching unit, where the first network connection is associated with a first communication port of the first network switching unit; suspending the first communication port from a link aggregation group (LAG), where the first communication port is associated with the LAG; and associating one or more first inter-chassis link (ICL) ports with the LAG. The first ICL ports are associated with a first ICL coupling the first network switching unit to a third network switching unit. The first network switching unit and the third network switching unit are peers. | 08-28-2014 |
20150023351 | METHOD AND APPARATUS FOR PERFORMING LINK AGGREGATION - A network node or corresponding method of performing link aggregation reduces a number of Content Addressable Memory (CAM) entries required to make a forwarding decision for a given ingress flow, reducing cost, size, and power consumption of the CAM and accompanying static RAM. In one embodiment, an ingress flow is mapped to an egress flow identifier. Subsequently, the egress flow identifier is mapped to a member of an aggregated group associated with an egress interface based on information available in a given ingress flow. Finally, the given ingress flow is forwarded to the member of the aggregated group associated with the egress interface. A hashing technique or two lookups may be used alone or in combination in mapping the ingress flow to the egress flow identifier to reduce CAM memory usage. | 01-22-2015 |
Patent application number | Description | Published |
20080288950 | Concurrent Management of Adaptive Programs - A method for concurrent management of adaptive programs is disclosed wherein changes in a set of modifiable references are initially identified. A list of uses of the changed references is next computed using records made in structures of the references. The list is next inserted into an elimination queue. Comparison is next made of each of the uses to the other uses to determine independence or dependence thereon. Determined dependent uses are eliminated and the preceding steps are repeated for all determined independent uses until all dependencies have been eliminated. | 11-20-2008 |
20090172294 | Method and apparatus for supporting scalable coherence on many-core products through restricted exposure - In one embodiment, a multi-core processor having cores each associated with a cache memory, can operate such that when a first core is to access data owned by a second core present in a cache line associated with the second core, responsive to a request from the first core, cache coherency state information associated with the cache line is not updated. A coherence engine associated with the processor may receive the data access request and determine that the data is of a memory page owned by the first core and convert the data access request to a non-cache coherent request. Other embodiments are described and claimed. | 07-02-2009 |
20100118041 | Shared virtual memory - Embodiments of the invention provide a programming model for CPU-GPU platforms. In particular, embodiments of the invention provide a uniform programming model for both integrated and discrete devices. The model also works uniformly for multiple GPU cards and hybrid GPU systems (discrete and integrated). This allows software vendors to write a single application stack and target it to all the different platforms. Additionally, embodiments of the invention provide a shared memory model between the CPU and GPU. Instead of sharing the entire virtual address space, only a part of the virtual address space needs to be shared. This allows efficient implementation in both discrete and integrated settings. | 05-13-2010 |
20100122264 | Language level support for shared virtual memory - Embodiments of the invention provide language support for CPU-GPU platforms. In one embodiment, code can be flexibly executed on both the CPU and GPU. CPU code can offload a kernel to the GPU. That kernel may in turn call preexisting libraries on the CPU, or make other calls into CPU functions. This allows an application to be built without requiring the entire call chain to be recompiled. Additionally, in one embodiment data may be shared seamlessly between CPU and GPU. This includes sharing objects that may have virtual functions. Embodiments thus ensure the right virtual function gets invoked on the CPU or the GPU if a virtual function is called by either the CPU or GPU. | 05-13-2010 |
20110238926 | Method And Apparatus For Supporting Scalable Coherence On Many-Core Products Through Restricted Exposure - In one embodiment, a multi-core processor having cores each associated with a cache memory, can operate such that when a first core is to access data owned by a second core present in a cache line associated with the second core, responsive to a request from the first core, cache coherency state information associated with the cache line is not updated. A coherence engine associated with the processor may receive the data access request and determine that the data is of a memory page owned by the first core and convert the data access request to a non-cache coherent request. Other embodiments are described and claimed. | 09-29-2011 |
20120159495 | NON-BLOCKING WAIT-FREE DATA-PARALLEL SCHEDULER - Methods, systems, and mediums are described for scheduling data parallel tasks onto multiple thread execution units of processing system. Embodiments of a lock-free queue structure and methods of operation are described to implement a method for scheduling fine-grained data-parallel tasks for execution in a computing system. The work of one of a plurality of worker threads is wait-free with respect to the other worker threads. Each node of the queue holds a reference to a task that may be concurrently performed by multiple thread execution units, but each on a different subset of data. Various embodiments relate to software-based scheduling of data-parallel tasks on a multi-threaded computing platform that does not perform such scheduling in hardware. Other embodiments are also described and claimed. | 06-21-2012 |
20130061240 | TWO WAY COMMUNICATION SUPPORT FOR HETEROGENOUS PROCESSORS OF A COMPUTER PLATFORM - A computer system may comprise a computer platform and input-output devices. The computer platform may include a plurality of heterogeneous processors comprising a central processing unit (CPU) and a graphics processing unit) GPU, for example. The GPU may be coupled to a GPU compiler and a GPU linker/loader and the CPU may be coupled to a CPU compiler and a CPU linker/loader. The user may create a shared object in an object oriented language and the shared object may include virtual functions. The shared object may be fine grain partitioned between the heterogeneous processors. The GPU compiler may allocate the shared object to the CPU and may create a first and a second enabling path to allow the GPU to invoke virtual functions of the shared object. Thus, the shared object that may include virtual functions may be shared seamlessly between the CPU and the GPU. | 03-07-2013 |
20130187936 | Language Level Support for Shared Virtual Memory - Embodiments of the invention provide language support for CPU-GPU platforms. In one embodiment, code can be flexibly executed on both the CPU and GPU. CPU code can offload a kernel to the GPU. That kernel may in turn call preexisting libraries on the CPU, or make other calls into CPU functions. This allows an application to be built without requiring the entire call chain to be recompiled. Additionally, in one embodiment data may be shared seamlessly between CPU and GPU. This includes sharing objects that may have virtual functions. Embodiments thus ensure the right virtual function gets invoked on the CPU or the GPU if a virtual function is called by either the CPU or GPU. | 07-25-2013 |
20140049550 | Shared Virtual Memory - Embodiments of the invention provide a programming model for CPU-GPU platforms. In particular, embodiments of the invention provide a uniform programming model for both integrated and discrete devices. The model also works uniformly for multiple GPU cards and hybrid GPU systems (discrete and integrated). This allows software vendors to write a single application stack and target it to all the different platforms. Additionally, embodiments of the invention provide a shared memory model between the CPU and GPU. Instead of sharing the entire virtual address space, only a part of the virtual address space needs to be shared. This allows efficient implementation in both discrete and integrated settings. | 02-20-2014 |
20140306972 | Language Level Support for Shared Virtual Memory - Embodiments of the invention provide language support for CPU-GPU platforms. In one embodiment, code can be flexibly executed on both the CPU and GPU. CPU code can offload a kernel to the GPU. That kernel may in turn call preexisting libraries on the CPU, or make other calls into CPU functions. This allows an application to be built without requiring the entire call chain to be recompiled. Additionally, in one embodiment data may be shared seamlessly between CPU and GPU. This includes sharing objects that may have virtual functions. Embodiments thus ensure the right virtual function gets invoked on the CPU or the GPU if a virtual function is called by either the CPU or GPU. | 10-16-2014 |
20140375662 | SHARED VIRTUAL MEMORY - Embodiments of the invention provide a programming model for CPU-GPU platforms. In particular, embodiments of the invention provide a uniform programming model for both integrated and discrete devices. The model also works uniformly for multiple GPU cards and hybrid GPU systems (discrete and integrated). This allows software vendors to write a single application stack and target it to all the different platforms. Additionally, embodiments of the invention provide a shared memory model between the CPU and GPU. Instead of sharing the entire virtual address space, only a part of the virtual address space needs to be shared. This allows efficient implementation in both discrete and integrated settings. | 12-25-2014 |
20150123978 | Shared Virtual Memory - Embodiments of the invention provide a programming model for CPU-GPU platforms. In particular, embodiments of the invention provide a uniform programming model for both integrated and discrete devices. The model also works uniformly for multiple GPU cards and hybrid GPU systems (discrete and integrated). This allows software vendors to write a single application stack and target it to all the different platforms. Additionally, embodiments of the invention provide a shared memory model between the CPU and GPU. Instead of sharing the entire virtual address space, only a part of the virtual address space needs to be shared. This allows efficient implementation in both discrete and integrated settings. | 05-07-2015 |
Patent application number | Description | Published |
20080280457 | METHOD OF IMPROVING INITIATION LAYER FOR LOW-K DIELECTRIC FILM BY DIGITAL LIQUID FLOW METER - A method for depositing a low dielectric constant film by flowing a oxidizing gas into a processing chamber, flowing an organosilicon compound from a bulk storage container through a digital liquid flow meter at an organosilicon flow rate to a vaporization injection valve, vaporizing the organosilicon compound and flowing the organosilicon compound and a carrier gas into the processing chamber, maintaining the organosilicon flow rate to deposit an initiation layer, flowing a porogen compound from a bulk storage container through a digital liquid flow meter at a porogen flow rate to a vaporization injection valve, vaporizing the porogen compound and flowing the porogen compound and a carrier gas into the processing chamber, increasing the organosilicon flow rate and the porogen flow rate while depositing a transition layer, and maintaining a second organosilicon flow rate and a second porogen flow rate to deposit a porogen containing organosilicate dielectric layer. | 11-13-2008 |
20090011148 | Methods and apparatuses promoting adhesion of dielectric barrier film to copper - Adhesion between a copper metallization layer and a dielectric barrier film may be promoted by stabilizing a flow of a silicon-containing precursor in a divert line leading to the chamber exhaust. The stabilized gas flow is then introduced to the processing chamber to precisely form a silicide layer over the copper. This silicidation step creates a network of strong Cu—Si bonds that prevent delamination of the barrier layer, while not substantially altering the sheet resistance and other electrical properties of the resulting metallization structure. | 01-08-2009 |
20100087062 | HIGH TEMPERATURE BD DEVELOPMENT FOR MEMORY APPLICATIONS - A method and apparatus for depositing organosilicate dielectric layers having good adhesion properties and low dielectric constant. Embodiments are described in which layers are deposited at low temperature and at high temperature. The low temperature layers are generally post-treated, whereas the high temperature layers need no post treating. Adhesion of the layers is promoted by use of an initiation layer. | 04-08-2010 |
20110136327 | HIGH MOBILITY MONOLITHIC P-I-N DIODES - Methods of forming high-current density vertical p-i-n diodes on a substrate are described. The methods include the steps of concurrently combining a group-IV-element-containing precursor with a sequential exposure to an n-type dopant precursor and a p-type dopant precursor in either order. An intrinsic layer is deposited between the n-type and p-type layers by reducing or eliminating the flow of the dopant precursors while flowing the group-IV-element-containing precursor. The substrate may reside in the same processing chamber during the deposition of each of the n-type layer, intrinsic layer and p-type layer and the substrate is not exposed to atmosphere between the depositions of adjacent layers. | 06-09-2011 |
20110223765 | SILICON NITRIDE PASSIVATION LAYER FOR COVERING HIGH ASPECT RATIO FEATURES - A method of forming a passivation layer comprising silicon nitride on features of a substrate is described. In a first stage of the deposition method, a dielectric deposition gas, comprising a silicon-containing gas and a nitrogen-containing gas, is introduced into the process zone and energized to deposit a silicon nitride layer. In a second stage, a treatment gas, having a different composition than that of the dielectric deposition gas, is introduced into the process zone and energized to treat the silicon nitride layer. The first and second stages can be performed a plurality of times. | 09-15-2011 |
20120164827 | FABRICATION OF THROUGH-SILICON VIAS ON SILICON WAFERS - A through-silicon via fabrication method comprises forming a substrate by bonding the front surface of a silicon plate to a carrier using an adhesive layer therebetween to expose the back surface of the silicon plate. A silicon nitride passivation layer is deposited on the exposed back surface of the silicon plate of the substrate. A plurality of through holes are etched in the silicon plate, the through holes comprising sidewalls and bottom walls. A metallic conductor is deposited in the through holes to form a plurality of through-silicon vias. | 06-28-2012 |
20120164829 | FABRICATION OF THROUGH-SILICON VIAS ON SILICON WAFERS - A through-silicon via fabrication method includes etching a plurality of through holes in a silicon plate. An oxide liner is deposited on the surface of the silicon plate and on the sidewalls and bottom wall of the through holes. A metallic conductor is then deposited in the through holes. In another version, which may be used concurrently with the oxide liner, a silicon nitride passivation layer is deposited on the exposed back surface of the silicon plate of the substrate. | 06-28-2012 |
20130161629 | ZERO SHRINKAGE SMOOTH INTERFACE OXY-NITRIDE AND OXY-AMORPHOUS-SILICON STACKS FOR 3D MEMORY VERTICAL GATE APPLICATION - Methods are provided for depositing a stack of film layers for use in vertical gates for 3D memory devices, by depositing a sacrificial nitride film layer at a sacrificial film deposition temperature greater than about 550° C.; depositing an oxide film layer over the nitride film layer, at an oxide deposition temperature of about 600° C. or greater; repeating the above steps to deposit a film stack having alternating layers of the sacrificial films and the oxide films; forming a plurality of holes in the film stack; and depositing polysilicon in the plurality of holes in the film stack at a polysilicon process temperature of about 700° C. or greater, wherein the sacrificial film layers and the oxide film layers experience near zero shrinkage during the polysilicon deposition. Flash drive memory devices may also be made by these methods. | 06-27-2013 |
20130230986 | ADHESION IMPROVEMENT FOR LOW K DIELECTRICS TO CONDUCTIVE MATERIALS - Methods are provided for processing a substrate for depositing an adhesion layer between a conductive material and a dielectric layer. In one aspect, the invention provides a method for processing a substrate including positioning a substrate having a conductive material disposed thereon, introducing a reducing compound or a silicon based compound, exposing the conductive material to the reducing compound or the silicon based compound, and depositing a silicon carbide layer without breaking vacuum. | 09-05-2013 |
20140118751 | PECVD PROCESS - A method of processing a substrate according to a PECVD process is described. Temperature profile of the substrate is adjusted to change deposition rate profile across the substrate. Plasma density profile is adjusted to change deposition rate profile across the substrate. Chamber surfaces exposed to the plasma are heated to improve plasma density uniformity and reduce formation of low quality deposits on chamber surfaces. In situ metrology may be used to monitor progress of a deposition process and trigger control actions involving substrate temperature profile, plasma density profile, pressure, temperature, and flow of reactants. | 05-01-2014 |
20140272184 | METHODS FOR MAINTAINING CLEAN ETCH RATE AND REDUCING PARTICULATE CONTAMINATION WITH PECVD OF AMORPHOUS SILICON FILIMS - Methods for maintaining clean etch rate and reducing particulate contamination with PECVD of amorphous silicon films are provided. The method comprises cleaning a processing chamber with a plasma comprising a cleaning gas, exposing at least a portion of the interior surfaces and components of the processing chamber to an oxidation gas and a nitration gas in the presence of a plasma and depositing a bi-layer seasoning layer on the interior surfaces and components of the processing chamber. | 09-18-2014 |
20140287593 | HIGH THROUGHPUT MULTI-LAYER STACK DEPOSITION - Methods and apparatus for high rate formation of multi-layer stacks on semiconductor substrate is provided. A chamber for forming such stacks at high rates includes a first precursor line and a second precursor line. The first precursor line is coupled to a first diverter, which is coupled to a gas inlet in a lid assembly of the chamber. The second precursor line is coupled to a second diverter, which is also coupled to the gas inlet. The first diverter is also coupled to a first divert line, and the second diverter is coupled to a second divert line. Each of the first and second divert lines is coupled to a divert exhaust system. A chamber exhaust system is coupled to the chamber. The diverters are typically located close to the lid assembly. | 09-25-2014 |
20150206757 | DIELECTRIC-METAL STACK FOR 3D FLASH MEMORY APPLICATION - A method is provided for forming a stack of film layers for use in 3D memory devices. The method starts with providing a substrate in a processing chamber of a deposition reactor. Then one or more process gases suitable for forming a dielectric layer are supplied into the processing chamber of the deposition reactor forming a dielectric layer on the substrate. Then one or more process gases suitable for forming a metallic layer are supplied into the processing chamber of the deposition reactor forming a metallic layer on the dielectric layer. Then one or more process gases suitable for forming a metallic nitride adhesion layer are supplied into the processing chamber of the deposition reactor forming a metallic nitride adhesion layer on the metallic layer. The sequence is then repeated to form a desired number of layers. | 07-23-2015 |
20150226540 | PECVD APPARATUS AND PROCESS - Apparatus and method of processing a substrate according to a PECVD process is described. Temperature profile of the substrate is adjusted to change deposition rate profile across the substrate. Plasma density profile is adjusted to change deposition rate profile across the substrate. Chamber surfaces exposed to the plasma are heated to improve plasma density uniformity and reduce formation of low quality deposits on chamber surfaces. In situ metrology may be used to monitor progress of a deposition process and trigger control actions involving substrate temperature profile, plasma density profile, pressure, temperature, and flow of reactants. | 08-13-2015 |
Patent application number | Description | Published |
20120014253 | MITIGATING THE EFFECTS OF CONGESTED INTERFACES ON A FABRIC - Techniques are provided for mitigating the effects of slow or no drain devices on a fabric. One or more of the described embodiments can be used alone or in combination to address problems associated with inter-switch link blocking and to address the situation where flows which are not associated with slow/no drain devices suffer the negative impacts of slow or no drain devices on a fabric. | 01-19-2012 |
20120294305 | Frame Handling Within Multi-Stage Switching Fabrics - Various techniques can be used to handle frames within multi-stage switching fabric. For example, in one method, a frame and an associated frame header are received at a switching fabric stage. The associated frame header includes a first field and a second field. The method selects one or more fabric points of exit within the switching fabric stage, based on the second field. The first field is used to select one or more other fabric points of exit within another switching fabric stage, and thus two different fields within the associated frame header specify fabric points of exit. The method then sends the frame to the selected fabric points of exit within the switching fabric stage. | 11-22-2012 |
20130258850 | CONGESTION ALLEVIATION IN NETWORK SYSTEMS - In one embodiment, a method is provided for alleviating congestion in a network system. In this method, the receipt of data packets destined for a destination apparatus is detected. Flow control signals are also received with each flow control signal corresponding to a data packet. Various time periods are tracked with each time period being between the detection of the receipt of a data packet and the receipt of tracked corresponding flow control signal. An average of the time periods is calculated and this average is compared to a threshold. One or more data packets are dropped in reference to the comparison. | 10-03-2013 |
20130258851 | MITIGATION OF CONGESTION DUE TO STUCK PORTS IN NETWORK SYSTEMS - In one embodiment, a method is provided for controlling congestion in a network system. In this method, receipt of a data packet that is destined for a destination switching apparatus is detected. Subsequent to the detection of the data packet, a time that has elapsed while flow control is implemented by the destination switching apparatus is tracked. The data packet is dropped based on the elapsed time exceeding a predefined time period. | 10-03-2013 |
20140086054 | MITIGATING THE EFFECTS OF CONGESTED INTERFACES ON A FABRIC - Techniques are provided for mitigating the effects of slow or no drain devices on a fabric. One or more of the described embodiments can be used alone or in combination to address problems associated with inter-switch link blocking and to address the situation where flows which are not associated with slow/no drain devices suffer the negative impacts of slow or no drain devices on a fabric. | 03-27-2014 |
20140280499 | DISTRIBUTED NETWORK SERVICES - Providing a distributed network service includes: receiving network traffic at a first physical device; and executing a service engine to participate in the distributed network service. The distributed network service is provided to at least the first target application instance executing in a first VM on the first physical device, and a second target application instance executing in a second VM on a second physical device; and a shared state of the distributed network service is maintained with respect to the first physical device and the second physical device. | 09-18-2014 |
20140282526 | MANAGING AND CONTROLLING A DISTRIBUTED NETWORK SERVICE PLATFORM - A distributed network service platform comprises: a logical data plane configured to process packets that are received by a plurality of physical devices, transmitted by the plurality of physical devices, or both, the logical data plane being physically distributed on the plurality of physical devices; and a logical control plane configured to manage and control the logical data plane, the logical control plane comprising one or more physical control planes operating on one or more physical devices. | 09-18-2014 |
20150149635 | METHOD AND SYSTEM FOR DISTRIBUTED LOAD BALANCING - Load balancing includes receiving, from a client, a connection request to establish a connection with a server; determining load balancing state information based at least in part on the connection request; synchronizing the determined load balancing state information across a plurality of service engines, including to invoke an atomic read-miss-create (RMC) function on a distributed data store service; and distributing the connection to a selected server among a plurality of servers according to a result of the RMC function. | 05-28-2015 |
Patent application number | Description | Published |
20100309820 | Preventing loss of network traffic due to inconsistent configurations within the network - In one embodiment, a method includes receiving at a processor at a node, notification of an error in a VLAN to topology mapping at the node, receiving a multi-destination packet from the VLAN in the VLAN to topology mapping, the multi-destination packet including a tree identifier associated with one of the topologies, and transmitting the multi-destination packet to all forwarding ports at the node in an unpruned tree corresponding to the tree identifier contained in the multi-destination packet. An apparatus is also disclosed. | 12-09-2010 |
20110038257 | METHOD AND APPARATUS FOR SEQUENCING OPERATIONS FOR AN INCOMING INTERFACE CHECK IN DATA CENTER ETHERNET - In one embodiment, a method includes obtaining an indication that a state associated with a node is to be changed and preventing data from being received on a first link. The method also includes updating at least one selected from a group including an incoming interface check (IIC) table and an outgoing interface (OIF) table to reflect the state. The state indicates that a second link is to be activated. Finally, the method includes allowing the data to be received on the second link after updating either or both the IIC table and the OIF table and after the timer duration for the NULL value for IIC has expired. | 02-17-2011 |
20110134802 | Determining A Routing Tree For Networks With Different Routing Protocols - According to one embodiment, a routing tree may be determined by facilitating communication of a first network and a second network. The first network comprises first switches and uses a first routing protocol, and the second network comprises second switches and uses a second routing protocol. The intersection of the first and second switches comprises gateway switches. A gateway switch enables creation of a second routing tree of the second routing protocol. The second routing tree has virtual links and a virtual root switch representing a first root switch of the first network. The second switches generate minimum link cost tunnels using information from the second routing protocol. A first routing tree of the first routing protocol is extended with the tunnels to merge the first routing tree and the second routing tree. | 06-09-2011 |
20110228670 | N_Port ID Virtualization node redundancy - In one embodiment, a method includes establishing a link between two N_Port Identifier Virtualization (NPIV) switches, the link having a high cost assigned thereto. The NPIV switches are in communication with a plurality of hosts through an N_Port Virtualization (NPV) device. The method further includes receiving at a first of the NPIV switches, an indication of a failure at a second of the NPIV switches, receiving data at the first NPIV switch, the data destined for one of the hosts associated with a domain of the second NPIV switch, and forwarding the data to the NPV device for delivery to the host, wherein a Fibre Channel Identifier (FCID) of the host is the same before and after the failure at the second NPIV switch. An apparatus is also disclosed. | 09-22-2011 |
20110273990 | Per-graph link cost assignment in layer 2 multipath networks - In one embodiment, a method includes assigning at a switch in a layer 2 multipath network, costs to a link in the network, each of the link costs associated with a different graph for forwarding traffic in the network, transmitting the link costs to other switches in the layer | 11-10-2011 |
20120027017 | MULTI-DESTINATION FORWARDING IN NETWORK CLOUDS WHICH INCLUDE EMULATED SWITCHES - Techniques are described which facilitate multi-destination forwarding in a Layer 2 Multipath (L2MP) network which includes an emulated switch. The emulated switch may correspond to two or more underlying peer link switches in the L2MP network, in which each of the peer link switches is linked to a Classical Ethernet (CE) switch over a virtual port channel (vPC). Traffic received by one of the peer link switches over the vPC is automatically forwarded to the other peer link switch (or switches). Multi-destination frames originating from the L2MP network addressed to hosts within the CE network are sent over only one of the peer link switches. | 02-02-2012 |
Patent application number | Description | Published |
20130083660 | Per-Group ECMP for Multidestination Traffic in DCE/TRILL Networks - Consistent with embodiments of the present disclosure, systems and methods are disclosed for providing per-group ECMP for multidestination traffic in a DCE/TRILL network. Embodiments enable per-group load balancing of multidestination traffic in DCE/L2MP networks by creating a new IS-IS PDU to convey the affinity of the parent node for a given multicast group. For broadcast and unknown unicast flooded traffic, the load balancing may be done on a per-vlan basis. | 04-04-2013 |
20130315233 | LARGE DISTRIBUTED FABRIC-BASED SWITCH USING VIRTUAL SWITCHES AND VIRTUAL CONTROLLERS - In one embodiment, a system includes at least one distributed line card (DLC) in electrical communication with at least one switch fabric coupler (SFC) and a meta-controller in electrical communication with the at least one DLC, the meta-controller including a processor for executing logic, logic configured for creating more than one virtual switch and a corresponding virtual control domain (VCD) associated with each virtual switch, wherein each virtual switch includes a plurality of physical ports of one or more of the at least one DLC, and logic configured for autonomously controlling the virtual switches, wherein the system appears to a device external of the system as a single cell switched domain. Other systems are described according to more embodiments. | 11-28-2013 |
20130315234 | METHOD FOR CONTROLLING LARGE DISTRIBUTED FABRIC-BASED SWITCH USING VIRTUAL SWITCHES AND VIRTUAL CONTROLLERS - In one embodiment, a method includes splitting a plurality of physical ports distributed across multiple distributed line cards (DLCs) into a plurality of virtual switches, wherein each virtual switch comprises ports of one or more DLC, creating a virtual control domain (VCD) associated with each virtual switch, and autonomously controlling the virtual switches, wherein each DLC is in electrical communication with at least one switch fabric coupler (SFC), and wherein the plurality of physical ports appear to external devices as being part of a single cell switched domain. Other systems and methods are described according to more embodiments. | 11-28-2013 |
20130329727 | SYSTEM AND METHOD FOR LAYER-2 MULTICAST MULTIPATHING - An example method is provided and includes a multicast data message from a data source, the message in a first virtual local area network and being associated with a multicast group. The method also includes calculating a hash value based on the virtual local area network, the data source, and the multicast group, determining a port for a designated router in a Layer-2 network based on the hash value, and switching the multicast data message to the port that was determined. | 12-12-2013 |
20140050091 | LOAD BALANCING OVERLAY NETWORK TRAFFIC USING A TEAMED SET OF NETWORK INTERFACE CARDS - A system includes a server including: logic adapted for receiving traffic from a virtual machine (VM), the traffic including at least one packet, logic adapted for hashing at least a portion of the at least one packet according to a hashing algorithm to obtain a hash value, and logic adapted for selecting an uplink based on the hash value; at least one accelerated network interface card (NIC), each accelerated NIC including: network ports including multiple Peripheral Component Interconnect express (PCIe) ports adapted for communicating with the server and a network, each network port including an uplink, logic adapted for encapsulating the at least one packet into an overlay-encapsulated packet, logic adapted for storing a media access control (MAC) address corresponding to the selected uplink as a source MAC (SMAC) address in an outer header of the overlay-encapsulated packet, and logic adapted for sending the overlay-encapsulated packet via the selected uplink. | 02-20-2014 |
20140153437 | Determining Multicast Root Nodes in a Telecommunications Network - In an embodiment, a method comprises: detecting a change in a multiple-switch configuration in a data communications network comprising a plurality of packet data switches configured as roots of multicast trees. In response to detecting that the multiple-switch configuration has changed, a first value, a second value and a third value representing limits on a number of multicast trees supported in the network and prioritization of the switches are retrieved. The method further comprises determining a type of the multiple-switch configuration change. In response to determining that the type indicates that a first switch was added to the multiple-switch configuration, using at least the first, second and third values, it is determined whether to configure the first switch as a first root in the multiple-switch configuration. The method is performed by one or more computing devices. | 06-05-2014 |
20140201733 | SCALABLE NETWORK OVERLAY VIRTUALIZATION USING CONVENTIONAL VIRTUAL SWITCHES - In one embodiment, a system includes a server running a virtualization platform, the virtualization platform including logic adapted for creating one or more virtual machines (VMs) and logic adapted for managing a virtual switch (vSwitch), a controller in communication with the server, the controller including logic adapted for assigning a media access control (MAC) address and a virtual local area network (VLAN) identifier (ID) to each of the one or more VMs, wherein a specific tenant to which the one or more VMs belongs is indicated using a tenant ID derived from the VLAN ID, the MAC address, or a combination thereof. Other systems, methods, and computer program products are also described according to more embodiments. | 07-17-2014 |
20150372917 | LOAD BALANCING OVERLAY NETWORK TRAFFIC USING A TEAMED SET OF NETWORK INTERFACE CARDS - In one embodiment, a server includes a processor and logic integrated with and/or executable by the processor. The logic is configured to hash at least a portion of a packet according to a hashing algorithm to obtain a hash value. The logic is also configured to select an uplink with which to forward the packet based on the hash value. In another embodiment, a method includes hashing at least a portion of at least one packet according to a hashing algorithm to obtain a hash value using an accelerated network interface card (NIC) of a server. The accelerated NIC is configured to provide overlay functionality. The method also includes selecting an uplink, based on the hash value, from a plurality of uplinks available to send traffic out to a network using the accelerated NIC. | 12-24-2015 |
20160080162 | ECMP Parent Group Selection for Multidestination Traffic in DCE/TRILL Networks - Consistent with embodiments of the present disclosure, systems and methods are disclosed for providing per-group ECMP for multidestination traffic in a DCE/TRILL network. Embodiments enable per-group load balancing of multidestination traffic in DCE/L2MP networks by creating a new IS-IS PDU to convey the affinity of the parent node for a given multicast group. For broadcast and unknown unicast flooded traffic, the load balancing may be done on a per-vlan basis. | 03-17-2016 |
Patent application number | Description | Published |
20130322331 | METHOD AND APPARATUS FOR EXCLUDING GUARANTEED BIT RATE TRAFFIC BEARERS FROM LTE UL FLOW CONTROL - A method, an apparatus, and a computer program product for wireless communication are provided. The apparatus is a UE. The UE transmits data packets. The UE determines to implement a flow control to reduce a transmission rate of the data packets. The UE determines whether the data packets include known or potential real-time data packets. The UE refrains from implementing the flow control to reduce the transmission rate of the known/potential real-time data packets when the data packets include known/potential real-time data packets. | 12-05-2013 |
20140086145 | OPTIMIZED HARQ RECOMBINING WITHIN TRANSMISSION WINDOW - Techniques for optimized HARQ recombining are provided. In one exemplary embodiment, a method for wireless communication comprises receiving a broadcast message to determine a timing of a transmission window, receiving at least one transmission within the transmission window, and determining whether the at least one transmission is successfully decoded. The method further comprises instructing a lower protocol layer to ignore remaining transmissions within the transmission window upon a determination that the at least one transmission is successfully decoded, wherein the remaining transmissions and the at least one transmission comprise duplicate copies of a message segment. | 03-27-2014 |
20140119247 | ADAPTIVE MAINTENANCE OF UPLINK TIME ALIGNMENT - A method for wireless communication is disclosed. A time alignment timer is started. It is determined when the time alignment timer will expire. A time alignment request subframe that is prior to the time alignment timer expiring is identified. Uplink time alignment is requested in the time alignment request subframe. | 05-01-2014 |
20140250456 | CONCURRENT SERVICE ALERTS DURING ACTIVE COMMUNICATION SESSIONS - Various aspects of the present disclosure for wireless communication may intelligently present a concurrent communication session to a user while the user's user equipment (UE) is streaming and/or presenting an original communication session. Example aspects may receive streaming communication content at a mobile device during an active communication session, present the streaming communication content at the mobile device, receive a concurrent service indication at the mobile device during the active communication session, and the concurrent service (or notification thereof) at the mobile device while concurrently streaming the communication content. | 09-04-2014 |
20140274038 | APPARATUS AND METHOD FOR PERFORMING BETTER PLMN SEARCH DURING CONNECTED MODE DRX - Disclosed are methods and apparatus for performing a better PLMN (BPLMN) search during connected mode DRX. In one aspect, a user equipment (UE) detects a trigger to begin a BPLMN search while in connected mode discontinuous reception (CDRX); determines whether it is configured for a manual BPLMN search; determine whether the CDRX duration is greater than or equal to the minimum amount of time required to perform the manual BPLMN search; and when determined that the CDRX duration is greater than or equal to the minimum amount of time required for the BPLMN search, performs the manual BPLMN search while in the CDRX. | 09-18-2014 |
20150094003 | SYSTEMS AND METHODS FOR TRANSMIT ANTENNA SWITCHING - A method for antenna switching is described. The method includes transmitting using a first antenna. The method also includes determining that a trigger occurs to switch to transmitting using a second antenna. The trigger is based on a combination of a physical uplink shared channel (PUSCH) maximum transmit power level (MTPL) counter and a physical uplink control channel (PUCCH) MTPL counter. The method further includes switching to transmitting using the second antenna based on the determination. | 04-02-2015 |
20150098341 | DEFENSE AGAINST FALSE DETECTION OF SEMI-PERSISTENT SCHEDULING (SPS) ACTIVATION OR RELEASE - Certain aspects of the present disclosure provide techniques for defending against false semi-persistent scheduling (SPS) activation detection and/or missed SPS release. According to certain aspects, a user equipment (UE) may detect one or more conditions for a semi-persistent scheduling (SPS) activation or release are met based on a downlink transmission, generate one or more metrics related to downlink transmission, and determine a valid SPS activation or release has occurred if the one or more metrics satisfy one or more criteria. According to certain aspects, a UE may determine a valid semi-persistent scheduling (SPS) activation has occurred, detect a number of PDSCH CRC failures, and implicitly declare an SPS release based on the detection. | 04-09-2015 |
20150131461 | METHODS AND APPARATUS FOR REDUCING MODEM POWER BASED ON A PRESENT STATE OF CHARGE OF BATTERY - Methods, systems, and devices are described for managing power of a user equipment (UE). A UE modem may determine the state of charge of the battery to determine that the battery is in one of two or more charge state levels, and may invoke one or more modem power saving modes based on the charge state level. Power saving modes may include, for example, reducing a number of available receive chains in a UE, initiating a time delay between one or more frequency scan requests performed by the UE, reducing a rate of neighbor search requests performed by the UE, providing a buffer status report (BSR) parameter that indicates a reduced amount of buffer data relative to an actual amount of buffer data for the UE, and/or adjusting a maximum transmit power level for an uplink channel. | 05-14-2015 |
20150163741 | DELAYING RADIO LINK CONTROL RETRANSMISSIONS - Certain aspects of the present disclosure relate to techniques and apparatus for a user equipment (UE) to delay RLC retransmissions (e.g., during off-durations, including CDRX off-durations). According to aspects of the present disclosure, a UE may delay triggering an RLC retransmission of an RLC PDU until after a next opportunity for the UE to receive an RLC ACK of the RLC PDU. By delaying RLC retransmissions, a UE may be prevented from waking up from one or more CDRX off-durations and using power associated with waking up from the one or more CDRX off-durations. | 06-11-2015 |
20150296391 | OPTIMIZED CELL ACQUISITION FOR LTE TDD SYSTEMS OR CSG/eMBMS CAPABLE UEs - Techniques for optimized cell acquisition for long term evolution (LTE) time division duplex (TDD) systems or closed subscriber group (CSG)/evolved Multicast Broadcast Multimedia Service (eMBMS) capable user equipments (UEs) are disclosed. A UE obtains configuration information, CSG capability information, and/or eMBMS capability information for a cell. The UE also stores the obtained information for the cell in a cell information database. In some aspects, the UE may employ the CSG and/or eMBMS capability information to select a weaker cell, during initial frequency scan at power up of the UE, based on cell capability. Additionally or alternatively, the UE may employ the configuration information to determine an initial mutual information (Mi) hypothesis value for physical hybrid-automatic repeat request (HART) indicator channel (PHICH) group mapping of the cell. Additionally or alternative, radio link failures and/or out of synch events may be predicted and proactive responses employed based on previously obtained solutions. | 10-15-2015 |
20150327156 | OPTIMIZATION OF MBSFN DECODING ON SCELLS WHEN THE PCELL AND SCELLS BELONG TO SAME MBSFN AREA - A mobile device determines that a primary cell (PCell) and a secondary cell (SCell) belong to a same multicast-broadcast single frequency network (MBSFN) area or that the PCell and the SCell can broadcast the same service. The mobile device determines a communication state of the SCell. If the SCell is active, the mobile device compares signal quality on the PCell and the SCell and selects to decode broadcast content on the PCell or the SCell. If the communication state indicating the SCell is deactivated or released, the mobile device switches to decode the broadcast content on the PCell. If the SCell is released and if the mobile device determines that the PCell is also released, then the mobile device switches to decode the broadcast content on the PCell and initiates a neighbor cell search. | 11-12-2015 |
20150351153 | ENHANCED PHYSICAL HARQ INDICATOR CHANNEL DECODING - Methods, systems, and devices are described for improving discontinuous reception (DRX) periods using enhanced physical HARQ indicator channel (PHICH) decoding. A user equipment (UE) may determine that an uplink (UL) retransmission (ReTx) is unnecessary based on the content of the original UL transmission. For example, the transmission may include media access control (MAC) layer padding rather than relevant application layer data. The UE may then identify a DRX sleep period that includes the subframe where the ReTx would take place. In some cases, the DRX sleep period may include a subframe where the UE would otherwise receive an acknowledgement message (AM) from a base station. The UE may then enter a DRX sleep state. In another example, the DRX sleep period is based on the content of a received AM. If the UE receives an ACK, the UL ReTx may be unnecessary. | 12-03-2015 |
Patent application number | Description | Published |
20080282020 | DETERMINATION OF SAMPLING CHARACTERISTICS BASED ON AVAILABLE MEMORY - A portion of data records of a full input data set are imported into memory of a computer system for processing by an executing application. The full input data set includes data records of a dimensionally-modeled fact collection. An amount of the data of the full input set to import is determined based on an amount of available memory of the computer system. The sampling characteristics for sampling the full input data set are determined based on the amount of the data that can be imported and on characteristics of the full input data set and application involved. The full input data set is then sampled and a portion of the records are imported into the memory of the computer system for processing. The sampling characteristics are determined such that analysis as a result of processing by the executing application of the sampled portion of the records imported is representative of the analysis that could otherwise be carried out on the full input data set, with a calculable statistical relevance. | 11-13-2008 |
20080288445 | Interactive aggregation of data on a scatter plot - A computer-implemented method, which comprises the following: aggregating a plurality of records in accordance with an aggregation specification, wherein the records are part of a dimensionally-modeled fact collection; graphically representing the records in un-aggregated form; graphically representing the records in aggregated form; and causing the graphical representation of the records to be switched between aggregated form and un-aggregated form based on a user indication. | 11-20-2008 |
20080288527 | User interface for graphically representing groups of data - A technique of operating a user interface that enables the user to graphically manipulate records of a dimensionally-modeled fact collection, which comprises the following: receiving a graphical selection of a subset from a set of data points, each data point representing at least one record of the dimensionally-modeled fact collection; receiving a graphical manipulation of the selected subset of data points; defining at least one data group using the selected subset of data points and based on the graphical manipulation, wherein each data group comprises between 0 to n records represented by the selected subset of data points, wherein n is the total number of data points in the set of data points; and graphically representing the at least one data group. Alternatively, the technique comprises the following: performing an operation on at least one data group as described above; and graphically representing a result of the operation. | 11-20-2008 |
20080294595 | VISUAL INTERFACE TO INDICATE CUSTOM BINNING OF ITEMS - Records representing items in a dimensionally-modeled fact collection are assigned to bins. A count-based portion of a user interface receives user bin assignment specification of the records based on user-specified counts of records. Actual counts for bin assignment are determined by constraining records having a same data value at a specified particular dimension to be within the same bin. A user-observable indication of the determined actual counts is provided. The user interface may include a value-based portion. The value-based portion of the user interface may be operated to receive user indication of bin assignment specification of records based on user-specified at least one value at the particular dimension. Determining actual counts includes reconciling the user indication of bin assignment specification in the count-based portion with the user indication of bin assignment specification in the value-based portion. | 11-27-2008 |
20080294671 | EXPORTING AGGREGATED AND UN-AGGREGATED DATA - A computer-implemented method, which comprises the following: receiving a graphical selection of a subset of data points from a set of data points, each data point representing at least one record of a dimensionally-modeled fact collection; and exporting information associated with the selected subset of data points. | 11-27-2008 |
Patent application number | Description | Published |
20150085834 | TIME DIVISION LONG TERM EVOLUTION (TD-LTE) FRAME STRUCTURE MODIFICATION - A method of wireless communication includes communicating with a base station using an extended special subframe. Communicating with the base station using the extended special subframe may be performed by disabling an uplink pilot time slot and an adjacent uplink subframe. | 03-26-2015 |
20150085840 | TIME DIVISION LONG TERM EVOLUTION (TD-LTE) FRAME STRUCTURE - A method of wireless communication includes communicating with a base station using a special subframe that extends a guard period over an uplink pilot time slot and one or more disabled, adjacent uplink subframes. The method also includes associating a control information subframe with a specific downlink subframe while accounting for both cell radius extension and loss of the one or more disabled, adjacent uplink subframes used to communicate the extended special subframe. | 03-26-2015 |
20150100318 | SYSTEMS AND METHODS FOR MITIGATING SPEECH SIGNAL QUALITY DEGRADATION - A method for decoding a speech signal is described. The method includes obtaining a packet. The method also includes obtaining a previous lag value. The method further includes limiting the previous lag value if the previous lag value is greater than a maximum lag threshold. The method additionally includes disallowing an adjustment to a number of synthesized peaks if a combination of the number of synthesized peaks and an estimated number of peaks is not valid. | 04-09-2015 |
20150181544 | ENHANCED RANDOM ACCESS PROCEDURE FOR AIR-TO-GROUND COMMUNICATIONS - Methods, systems, and devices are described for establishing a wireless communications link at an aircraft terminal (AT). An AT may determine a timing offset based on the propagation delay between the AT and a ground station, which in some cases may be more than 100 kilometers away. The AT may then transmit an initial access message to the ground station based on the determined timing offset. In some embodiments, the AT may receive an access response message from the ground station that includes a timing alignment value and adjust the timing offset based on this value. The AT may transmit a connection message to the ground station that includes AT location information. This information may be used by the ground station to facilitate beamforming. | 06-25-2015 |
20150334707 | TIMING ADVANCE TECHNIQUES FOR LARGE CELLS - Methods, systems, and devices are described for implementing timing advances in which a propagation delay may exceed a time period of a portion of a radio frame. In some examples, a transmitter may identify a timing advance indicating a time to initiate wireless uplink transmission of a subframe. Such a timing advance may compensate for a propagation delay between the transmitter and a receiver of the wireless uplink transmission. The timing advance may be applied as an integer component and a fractional component in relation to a duration of the subframe, to adjust the time to initiate the wireless uplink transmission of the subframe. The integer component may be used to adjust one or more subframe characteristics, and the fractional component that may be used to adjust the time to initiate the wireless uplink transmission of the subframe. | 11-19-2015 |
Patent application number | Description | Published |
20100011166 | Data Cache Virtual Hint Way Prediction, and Applications Thereof - A virtual hint based data cache way prediction scheme, and applications thereof. In an embodiment, a processor retrieves data from a data cache based on a virtual hint value or an alias way prediction value and forwards the data to dependent instructions before a physical address for the data is available. After the physical address is available, the physical address is compared to a physical address tag value for the forwarded data to verify that the forwarded data is the correct data. If the forwarded data is the correct data, a hit signal is generated. If the forwarded data is not the correct data, a miss signal is generated. Any instructions that operate on incorrect data are invalidated and/or replayed. | 01-14-2010 |
20150293853 | DATA CACHE VIRTUAL HINT WAY PREDICTION, AND APPLICATIONS THEREOF - A virtual hint based data cache way prediction scheme, and applications thereof. In an embodiment, a processor retrieves data from a data cache based on a virtual hint value or an alias way prediction value and forwards the data to dependent instructions before a physical address for the data is available. After the physical address is available, the physical address is compared to a physical address tag value for the forwarded data to verify that the forwarded data is the correct data. If the forwarded data is the correct data, a hit signal is generated. If the forwarded data is not the correct data, a miss signal is generated. Any instructions that operate on incorrect data are invalidated and/or replayed. | 10-15-2015 |