Patent application number | Description | Published |
20140095553 | TECHNIQUES FOR MOVING DATA FILES WITHOUT INTERRUPTING ACCESS - Techniques for moving data files without interrupting access are described. A first process moves a database file from a first location to a second location while the database file is accessible to one or more other processes for read or write operations. According to one technique, the first process communicates a move status and a copy range into the database file to one or more database server instances executing the one or more other processes. The one or more other processes then perform input/output (IO) operations on the database file based at least in part on the move status and the copy range communicated by the first process. | 04-03-2014 |
20140095805 | REMOTE-KEY BASED MEMORY BUFFER ACCESS CONTROL MECHANISM - A system and method implementing revocable secure remote keys is disclosed. A plurality of indexed base secrets is stored in a register of a coprocessor of a local node coupled with a local memory. When it is determined that a selected base secret expired, the base secret stored in the register based on the base secret index is changed, thereby invalidating remote keys generated based on the expired base secret. A remote key with validation data and a base secret index is received from a node requesting access to the local memory. A validation base secret is obtained from the register based on the base secret index. The coprocessor performs hardware validation on the validation data based on the validation base secret. Hardware validation fails if the base secret associated with the base secret index has been changed in the register of the selected coprocessor. | 04-03-2014 |
20140096145 | HARDWARE MESSAGE QUEUES FOR INTRA-CLUSTER COMMUNICATION - A method and apparatus for sending and receiving messages between nodes on a compute cluster is provided. Communication between nodes on a compute cluster, which do not share physical memory, is performed by passing messages over an I/O subsystem. Typically, each node includes a synchronization mechanism, a thread ready to receive connections, and other threads to process and reassemble messages. Frequently, a separate queue is maintained in memory for each node on the I/O subsystem sending messages to the receiving node. Such overhead increases latency and limits message throughput. Due to a specialized coprocessor running on each node, messages on an I/O subsystem are sent, received, authenticated, synchronized, and reassembled at a faster rate and with lower latency. Additionally, the memory structure used may reduce memory consumption by storing messages from multiple sources in the same memory structure, eliminating the need for per-source queues. | 04-03-2014 |
20140365734 | OBSERVATION OF DATA IN PERSISTENT MEMORY - Systems and methods for reliably using data storage media. Multiple processors are configured to access a persistent memory. For a given data block corresponding to a write access request from a first processor to the persistent memory, a cache controller prevents any read access of a copy of the given data block in an associated cache. The cache controller prevents any read access while detecting an acknowledgment that the given data block is stored in the persistent memory is not yet received. Until the acknowledgment is received, the cache controller allows write access of the copy of the given data block in the associated cache only for a thread in the first processor that originally sent the write access request. The cache controller invalidates any copy of the given data block in any cache levels below the associated cache. | 12-11-2014 |
Patent application number | Description | Published |
20090204753 | SYSTEM FOR REFRESHING CACHE RESULTS - A system and method for refreshing a cache based on query responses provided by a searching system in response to queries, includes providing a cache entry for each unique query, if space is available in the cache, and assigning a temperature value to each cache entry based on a frequency of occurrence of the corresponding query An age value is assigned to each cache entry based on a time of last refresh or creation of the corresponding query response. The age of the cache entries is periodically updated, and the temperature of a cache entry is updated when a corresponding query reoccurs. If system resources are available, the query response of a cache entry is refreshed based on the temperature and age of the cache entry. If resources are not available, the refreshing is limited. | 08-13-2009 |
20110138148 | Dynamic Data Storage Repartitioning - Embodiments of the present invention enable dynamic repartitioning of data storage in response to one or more triggers. In embodiments, a trigger may be a user-initiated action, a system-generated action, and/or an inference based on storage usage parameters. Applications of the present invention are its use in embodiments of a storage management system comprising a file system manager and a volume manager, where the placement of data into a partition (data storage region) may be specified by matching one or more disk region placement data attributes assigned to data with corresponding disk region attributes. In embodiments, dynamic repartitioning comprises adjustment of the location of the boundary between adjacent disk partitions and, if necessary, rebalancing of the data stored within the partitions by identifying mismatched data and relocating it to the partition with which it is associated. | 06-09-2011 |