Patent application number | Description | Published |
20080215450 | REMOTE PROVISIONING OF INFORMATION TECHNOLOGY - Remote provisioning of an IT network and/or associated services is provided. Hardware, software, service and/or expertise can be moved from on-premise to a remote location (e.g., central, distributed . . . ). Accordingly, at least a large degree computation can be moved to the center to exploit economies of scale, among other things. In such an architecture, computational resources (e.g., data storage, computation power, cache . . . ) can be pooled, and entities can subscribe to a particular level of resources related to a private entity IT network. | 09-04-2008 |
20090077573 | Execution Context Infrastructure - Described herein is an implementation for exposing an “execution context” to a logical execution flow of procedures as it executes. An “execution context” is a set of data and/or sub-procedures that might be useful at some point during a logical execution flow (to manage and control the execution flow and provide additional services to the execution flow) of computer-executable instructions though the often complex, intertwined, and interconnected conglomeration of procedures of software product(s) | 03-19-2009 |
20100070957 | System and Method for Sharing Objects Between Applications in a Virtual Runtime Environment - The techniques and mechanisms described herein are directed at sharing objects between applications executing in a virtual runtime environment. The techniques include compiling a cpu-independent file to create executable instructions that are associated with an application that is executable in a virtual runtime environment. A constructor associated with a shareable class defined within the cpu-independent file is executed to create an instance of a shareable object. The shareable object is then persisted in a manner such that other applications executing in the virtual runtime environment may access the shareable object. The technique may be performed during pre-JIT compilation or just-in-time compilation of the cpu-independent file. The execution of the constructor is performed at a restricted security level that prevents access to changeable information when generating static data for the static members of the shareable object. | 03-18-2010 |
Patent application number | Description | Published |
20080222659 | ABSTRACTING OPERATING ENVIRONMENT FROM OPERATING SYSTEM - The present invention extends to methods, systems, and computer program products for abstracting an operating environment from an operating system running in the operating environment. Within an operating environment, an operating environment abstraction layer abstracts and exposes operating environment resources to an operating system. Accordingly, appropriately configured operating environment abstraction layers provide the operating system with a uniform interface to available resources across a variety of different operating environments. Each operating environment abstraction layer and the operating system include adjustable algorithms that can be adjusted to appropriately provide services to requesting applications based on exposed resources of the operating environment. Abstraction layers can be configured to analyze and become fully aware of their operating environment, including identifying the presence of other abstraction layers. An operating system and corresponding abstraction layer can be run in flexible combinations of privileged and unprivileged processor modes. | 09-11-2008 |
20090187893 | COORDINATING CHORES IN A MULTIPROCESSING ENVIRONMENT - The coordination and execution of chores in a multiprocessing environment. The coordination of chores is accomplished utilizing a compiler generated correlation that relates blocks of code that execute chores and blocks of code in which the chore can be realized. By tracking the execution of the program and using the compiler-generated correlation, chores can be identified for the currently executing code. | 07-23-2009 |
20140195632 | IMMUTABLE SHARABLE ZERO-COPY DATA AND STREAMING - The environment and use of an immutable buffer. A computing entity acquires data or generates data and populates the data into the buffer, after which the buffer is classified as immutable. The classification protects the data populated within the immutable buffer from changing during the lifetime of the immutable buffer, and also protects the immutable buffer from having its physical address changed during the lifetime of the immutable buffer. As different computing entities consume data from the immutable buffer, they do so through views provided by a view providing entity. The immutable buffer architecture may also be used for streaming data in which each component of the streaming data uses an immutable buffer. Accordingly, different computing entities may view the immutable data differently without having to actually copy the data. | 07-10-2014 |
20140195739 | ZERO-COPY CACHING - Caching of an immutable buffer that has its data and address prevented from changing during the lifetime of the immutable buffer. A first computing entity maintains a cache of the immutable buffer and has a strong reference to the immutable buffer. So long as any entity has a strong reference to the immutable buffer, the immutable buffer is guaranteed to continue to exist for the duration of the strong reference. A second computing entity communicates with the first computing entity to obtain a strong reference to the immutable buffer and thereafter read data from the immutable buffer. Upon reading the data from the cache, the second computing entity demotes the strong reference to a weak reference to the immutable buffer. A weak reference to the immutable buffer does not guarantee that the immutable buffer will continue to exist for the duration of the weak reference. | 07-10-2014 |
20140195746 | DMA CHANNELS - Communicating between an application and a hardware device. A method includes an application writing data to host physical memory using an application view of the memory. The method further includes mapping the data in the physical memory to a hardware driver view, usable by a hardware driver, without needing to copy the data to a different physical storage location. The method further includes mapping the data to a hardware accessible view accessible by a hardware device without needing to copy the data to a different physical storage location | 07-10-2014 |
20140195834 | HIGH THROUGHPUT LOW LATENCY USER MODE DRIVERS IMPLEMENTED IN MANAGED CODE - Implementing a safe driver that can support high throughput and low latency devices. The method includes receiving a hardware message from a hardware device. The method further includes delivering the hardware message to one or more driver processes executing in user mode using a zero-copy to allow the one or more driver processes to support high throughput and low latency hardware devices. | 07-10-2014 |
20140196059 | CAPABILITY BASED DEVICE DRIVER FRAMEWORK - Enforcing limitations on hardware drivers. The method includes from a system kernel, assigning I/O resources to the system's root bus. From the root bus, the method further includes assigning a subset of the I/O resources to a device bus. Assigning a subset of the I/O resources to a device bus includes limiting the device bus to only be able to assign I/O resources that are assigned to it by the root bus. From the device bus, the method includes assigning I/O resources to a device through a device interface. | 07-10-2014 |