Patent application number | Description | Published |
20100135412 | MEDIA CODING FOR LOSS RECOVERY WITH REMOTELY PREDICTED DATA UNITS - An improved loss recovery method for coding streaming media classifies each data unit in the media stream as an independent data unit (I unit), a remotely predicted unit (R unit) or a predicted data unit (P unit). Each of these units is organized into independent segments having an I unit, multiple P units and R units interspersed among the P units. The beginning of each segment is the start of a random access point, while each R unit provides a loss recovery point that can be placed independently of the I unit. This approach separates the random access point from the loss recovery points provided by the R units, and makes the stream more impervious to data losses without substantially impacting coding efficiency. The most important data units are transmitted with the most reliability to ensure that the majority of the data received by the client is usable. The I units are the least sensitive to transmission losses because they are coded using only their own data. While they provide the best coding efficiency, the P units are the most sensitive to data loss because the loss of one P unit renders useless all of the P units that depend on it. The remotely predicted units are dependent on the I unit, or in an alternative implementation, on another R unit. | 06-03-2010 |
20130329779 | MEDIA CODING FOR LOSS RECOVERY WITH REMOTELY PREDICTED DATA UNITS - An improved loss recovery method for coding streaming media classifies each data unit in the media stream as an independent data unit (I unit), a remotely predicted unit (R unit) or a predicted data unit (P unit). Each of these units is organized into independent segments having an I unit, multiple P units and R units interspersed among the P units. The beginning of each segment is the start of a random access point, while each R unit provides a loss recovery point that can be placed independently of the I unit. This approach separates the random access point from the loss recovery points provided by the R units, and makes the stream more impervious to data losses without substantially impacting coding efficiency. The most important data units are transmitted with the most reliability to ensure that the majority of the data received by the client is usable. The I units are the least sensitive to transmission losses because they are coded using only their own data. While they provide the best coding efficiency, the P units are the most sensitive to data loss because the loss of one P unit renders useless all of the P units that depend on it. The remotely predicted units are dependent on the I unit, or in an alternative implementation, on another R unit. | 12-12-2013 |
Patent application number | Description | Published |
20090003387 | Synchronization Between Connection Manager and Extension Components - This disclosure describes synchronization between a connection manager and extension components during terminal service starts, stops, and restarts. The synchronization occurs by application programming interfaces as mechanisms to notify extension components that a change of state is about to occur. The extension components take appropriate action steps, such as saving session information for a connection pertaining to a connection stop and restoring connection session information when the terminal service is restarted. Furthermore, the extension components may implement own caching mechanisms and selectively perform lazy restore on save data as necessary. As a result, experience for the user is enhanced by not losing any data when the terminal service is stopped and restarted. Also, administrators may perform a patch on terminal service binaries without waiting for all users to log off or without rebooting the system. | 01-01-2009 |
20090006537 | Virtual Desktop Integration with Terminal Services - An integration system is disclosed that provides a virtual desktop integration with terminal services. A client computer is connected to one the virtual desktops operating in a server. The client computer examines information contained in a remote desktop protocol (RDP) compliant packet supplied by the server. The client computer connects to one of the many virtual desktops based on information. Use of the information enables integration of the virtual desktop with the existing terminal session deployment model. Client devices can establish a session using a single network name and can be appropriately directed to either a virtual desktop or terminal session. | 01-01-2009 |
20090100448 | Remote Auto Provisioning and Publication of Applications - Techniques for auto provisioning and publication of applications are described. The auto provisioning process enables automatic installation of applications having their respective file packages pre-stored in the server. Subsequently, publication of these applications is conducted automatically. By employing such an automatic process, the latency involved in provisioning and publication of remote applications is reduced, thereby making the process more efficient. | 04-16-2009 |
20090113058 | Terminal server draining - Techniques are described for managing and controlling connections of server sessions, including systems and methods for managing and controlling connections to a server computer. Connections are monitored to a server computer depending on a “drain state” of the server computer. If the server is in drain mode, requests to initiate a fresh session with the server computer are rejected, or deferred until the server computer comes out of the drain state. | 04-30-2009 |
20090241176 | LOAD BALANCING IN SERVER COMPUTER SYSTEMS - Systems and methods to implement load balancing of connections to a server computer in a server collection are described. The server collection receives connection requests from remote clients over a network. A session broker evaluates one or more load parameters of the server computers in the server collection and, based on those load parameters, determines load associated with each server computer. The session broker redirects the connection requests to the server computer which has a lesser load. | 09-24-2009 |
20110018883 | HYBRID REMOTE SESSIONS - High Fidelity remoting can be enabled by loading a hybrid remote session in a computer system. The hybrid remote session can include components loaded in a console session and components loaded in a remote session. | 01-27-2011 |
20130067451 | APPLICATION DEPLOYMENT AND REGISTRATION IN A MULTI-USER SYSTEM - Embodiments are directed to efficiently provisioning applications in a multi-user environment. In one scenario, an administrator computer system pre-installs at least a portion of a software application on a data image, where the software application is to be installed on a client computer system upon receiving a notification from a computer administrator. The administrator computer system receives an input from the computer administrator indicating that a specified software application pre-installed on the client computer system's data image is to be installed on the client computer system. The administrator computer system then publishes a notification to the client computer system indicating that a specified software application pre-installed on the client computer system's data image is to be installed. Upon receiving the notification, the software application is installed on the client computer system from the data image. | 03-14-2013 |
20130114624 | MINIMUM NETWORK BANDWIDTH IN MULTI-USER SYSTEM - The management of bandwidth utilization in a network system that has multiple users. The system identifies the current set of users that are accessing the network by tracking which users are newly accessing the network, which users have ceased accessing the network, and which users continue to access the network over a period of time. The system then guarantees that these identified users will have a certain minimum bandwidth over the period of time. The system then enforces the guaranteed minimum bandwidth for each of the users so that they have no less than the corresponding guaranteed minimum bandwidth. Thus, even if the network bandwidth utilization would be saturated if all requests were satisfied, the system may continue operating while provided some guaranteed level of bandwidth to each user. | 05-09-2013 |
20130117336 | DATA CHANNEL UTILIZATION MANAGEMENT IN MULTI-USER SYSTEM - The management of data storage channel utilization in a computing system that has multiple users. The system receives file-level requests from requesters and then creates a history for each requester. Upon evaluating the history of each requester, the system determines whether to delay the file-level requests from entering the file system stack based on the result of the evaluation. The system delays one or more of the file-level requests if the history of the corresponding requester meets one or more criteria. If the history of the corresponding requester does not meet the criteria, the system allows the file-level requests to be passed to the file system stack without being delayed. | 05-09-2013 |
Patent application number | Description | Published |
20090112949 | COMPRESSED STORAGE MANAGEMENT - In accordance with one or more aspects, compressed storage management in a system includes determining which of multiple data units stored in an uncompressed portion of the storage resource are to be compressed and stored in a compressed portion of the storage resource. The management can further include returning one or more regions of the compressed portion for use in the uncompressed portion in response to storage resource pressure in the system, as well as compacting regions in the compressed portion to fill empty gaps in the compressed portion. | 04-30-2009 |
20100332721 | OPERATING SYSTEM VIRTUAL MEMORY MANAGEMENT FOR HARDWARE TRANSACTIONAL MEMORY - Operating system virtual memory management for hardware transactional memory. A method may be performed in a computing environment where an application running on a first hardware thread has been in a hardware transaction, with transactional memory hardware state in cache entries correlated by memory hardware when data is read from or written to data cache entries. The data cache entries are correlated to physical addresses in a first physical page mapped from a first virtual page in a virtual memory page table. The method includes an operating system deciding to unmap the first virtual page. As a result, the operating system removes the mapping of the first virtual page to the first physical page from the virtual memory page table. As a result, the operating system performs an action to discard transactional memory hardware state for at least the first physical page. Embodiments may further suspend hardware transactions in kernel mode. Embodiments may further perform soft page fault handling without aborting a hardware transaction, resuming the hardware transaction upon return to user mode, and even successfully committing the hardware transaction. | 12-30-2010 |
20110113432 | COMPRESSED STORAGE MANAGEMENT - Compressed storage management includes assigning a selection priority and a priority level to multiple data units stored in an uncompressed portion of a storage resource. The management can further include compressing data units and storing the compressed data units in a compressed portion of the storage resource. The data units in the compressed portion are stored in regions, which each store data units having the same selection priority or the same selection priority level. | 05-12-2011 |
20110145552 | Handling Operating System (OS) Transitions In An Unbounded Transactional Memory (UTM) Mode - In one embodiment, the present invention includes a method for receiving control in a kernel mode via a ring transition from a user thread during execution of an unbounded transactional memory (UTM) transaction, updating a state of a transaction status register (TSR) associated with the user thread and storing the TSR with a context of the user thread, and later restoring the context during a transition from the kernel mode to the user thread. In this way, the UTM transaction may continue on resumption of the user thread. Other embodiments are described and claimed. | 06-16-2011 |
20120137167 | SYSTEMATIC MITIGATION OF MEMORY ERRORS - A system and method for mitigating memory errors in a computer system. Faulty memory is identified and tested by a memory manager of an operating system. The memory manager may perform diagnostic tests while the operating system is executing on the computer system. Regions of memory that are being used by software components of the computer system may also be tested. The memory manager maintains a stored information about faulty memory regions. Regions are added to the stored information when they are determined to be faulty by a diagnostic test tool. Memory regions are allocated to software components by the memory manager after checking the stored information about faulty memory regions. This ensures a faulty memory region is never allocated to a software component of the computer system. | 05-31-2012 |
20120233438 | PAGEFILE RESERVATIONS - A system and method for maintaining a pagefile of a computer system using a technique of reserving portions of the pagefile for related memory pages. Pages near one another in a virtual memory space often store related information and it is therefore beneficial to ensure that they are stored near each other in the pagefile. This increases the speed of reading data out of the pagefile because total seek time of a disk drive that stores the pagefile may decrease when adjacent pages in a virtual memory address space are read back from the disk drive. By implementing a reservation system that allows related pages to be stored adjacent to one another, the efficiency of memory management to of the computer system is increased. | 09-13-2012 |
20120246436 | COMBINING MEMORY PAGES HAVING IDENTICAL CONTENT - In a device having a memory accessed as multiple pages, two or more pages of the multiple pages having identical content are identified. While the two or more pages are being identified, other processes running in the device are allowed to use the two or more pages, including being allowed to change cache attributes of each of the two or more pages. The two or more pages are combined into a single combined page (e.g., a newly allocated page of the multiple pages), and a process page record having multiple entries pointing to the multiple pages is updated so that entries that previously pointed to one of the two or more pages instead point to the single page. | 09-27-2012 |
20120284485 | OPERATING SYSTEM VIRTUAL MEMORY MANAGEMENT FOR HARDWARE TRANSACTIONAL MEMORY - Operating system virtual memory management for hardware transactional memory. A system includes an operating system deciding to unmap a first virtual page. As a result, the operating system removes the mapping of the first virtual page to the first physical page from the virtual memory page table. As a result, the operating system performs an action to discard transactional memory hardware state for at least the first physical page. Embodiments may further suspend hardware transactions in kernel mode. Embodiments may further perform soft page fault handling without aborting a hardware transaction, resuming the hardware transaction upon return to user mode, and even successfully committing the hardware transaction. | 11-08-2012 |
20130132681 | TEMPORAL STANDBY LIST - In one embodiment, a memory management system temporarily maintains a memory page at an artificially high priority level | 05-23-2013 |
20130159662 | Working Set Swapping Using a Sequentially Ordered Swap File - Techniques described enable efficient swapping of memory pages to and from a working set of pages for a process through the use of large writes and reads of pages to and from sequentially ordered locations in secondary storage. When writing pages from a working set of a process into secondary storage, the pages may be written into reserved, contiguous locations in a dedicated swap file according to a virtual address order or other order. Such writing into sequentially ordered locations enables reading in of clusters of pages in large, sequential blocks of memory, providing for more efficient read operations to return pages to physical memory. | 06-20-2013 |
20140351552 | WORKING SET SWAPPING USING A SEQUENTIALLY ORDERED SWAP FILE - Techniques described enable efficient swapping of memory pages to and from a working set of pages for a process through the use of large writes and reads of pages to and from sequentially ordered locations in secondary storage. When writing pages from a working set of a process into secondary storage, the pages may be written into reserved, contiguous locations in a dedicated swap file according to a virtual address order or other order. Such writing into sequentially ordered locations enables reading in of clusters of pages in large, sequential blocks of memory, providing for more efficient read operations to return pages to physical memory. | 11-27-2014 |
20150039869 | Handling Operating System (Os) Transitions In An Unbounded Transactional Memory (Utm) Mode - In one embodiment, the present invention includes a method for receiving control in a kernel mode via a ring transition from a user thread during execution of an unbounded transactional memory (UTM) transaction, updating a state of a transaction status register (TSR) associated with the user thread and storing the TSR with a context of the user thread, and later restoring the context during a transition from the kernel mode to the user thread. In this way, the UTM transaction may continue on resumption of the user thread. Other embodiments are described and claimed. | 02-05-2015 |
20150324299 | TEMPORAL STANDBY LIST - In one embodiment, a memory management system temporarily maintains a memory page at an artificially high priority level. The memory management system may assign an initial priority level to a memory page in a page priority list. The memory management system may change the memory page to a target priority level in the page priority list after a protection period has expired. | 11-12-2015 |