Class / Patent application number | Description | Number of patent applications / Date published |
711160000 | Least recently used (LRU) | 46 |
20080270715 | Life Moment Tagging and Storage - A secure memory device and method for obtaining and securely storing information relating to a life moment is disclosed. In the method, a parameter is received and inputted in a search heuristic. A search is made for the information according to the search heuristic and, upon finding the information, metadata is appended to the information. The information and metadata is then stored in a secure memory location. The secure memory location has a housing fabricated to withstand a predetermined stress, a detachable connection to a computer and a memory that stores the information and protects it from unauthorized deletion. In some embodiments, the stored information may be selectively deleted in a safe and controlled manner. | 10-30-2008 |
20080288731 | Request arbitration device and memory controller - A bus arbiter receives requests of initiators, and internally includes a page hit/miss determining unit with permissible determining function, a bank open/close determining unit with permissible determining function, and an LRU unit with permissible determining function. Regarding the priority of the request arbitration on the requests, the bank priority on the SDRAM is determined in the order of page hit, bank open, and LRU. Furthermore, each determining unit internally includes a permissible time determining unit, and processes, at top priority, the request of the initiator which the corresponding permissible time is below the count threshold value in the priority processing of the determining unit. | 11-20-2008 |
20080307174 | Dual Use Memory Management Library - A dual-use library that is able to handle calls from programs requiring either reference count or garbage collected memory management is described. This capability may be provided by introducing a new assignment routine, assign( ), and instrumenting the reference count routines responsible for updating an object's reference count—e.g., addReference( ) and removeReference( ) routines. The assign( ), addReferenc( ) and removeReference( ) routines determine, at runtime, which memory management scheme is appropriate and execute the appropriate instructions (i.e., reference count or garbage collection specific instructions). The described dual-use library provides equivalent functionality as prior art two library implementations, but with a significantly lower memory footprint. | 12-11-2008 |
20080320256 | LRU control apparatus, LRU control method, and computer program product - To reduce the number of bits required for LRU control when the number of target entries is large, and achieve complete LRU control. Each time an entry is used, an ID of the used entry is stored to configure LRU information so that storage data | 12-25-2008 |
20090089520 | Hardware acceleration of strongly atomic software transactional memory - In accordance with some embodiments, software transactional memory may be used for both managed and unmanaged environments. If a cache line is resident in a cache and this is not the first time that the cache line has been read since the last write, then the data may be read directly from the cache line, improving performance. Otherwise, a normal read may be utilized to read the information. Similarly, write performance can be accelerated in some instances to improve performance. | 04-02-2009 |
20090177854 | METHODS, SYSTEMS, AND COMPUTER PROGRAM PRODUCTS FOR PREEMPTIVE PAGE EVICTION - A method, system, and computer program product for preemptive page eviction in a computer system are provided. The method includes identifying a region in an input file for preemptive page eviction, where the identified region is infrequently accessed relative to other regions of the input file. The method also includes generating an output file from the input file, where the identified region is flagged as a page for preemptive page eviction in the output file. The method further includes loading the output file to a memory hierarchy including a faster level of memory and a slower level of memory, wherein the flagged page is preemptively written to the slower level of memory. | 07-09-2009 |
20090193205 | DATA NETWORK AND A METHOD OF REGENERATION OF THE RECORDING STATE OF DIGITAL DATA IN A DATA NETWORK - A method of regeneration of a recording state of digital data stored in a node of a data network, the method including the steps of classifying files stored in the node, periodically writing a digital file from the node to a temporary memory, the temporary memory being a component of said node, and writing the digital file from the temporary memory to the same node. | 07-30-2009 |
20090276588 | Free space utilization in tiered storage systems - Embodiments of the invention include first storage mediums having first storage characteristics for making up a first pool of capacity of a first tier of storage, and second storage mediums having second storage characteristics for making up a second pool of capacity of a second tier of storage. Free capacity of the first and second pools is shared between the first and second tiers of storage. When the first pool has an amount of free capacity available over a reserved amount of free capacity reserved for first tier data, a first quantity of second tier data is moved from the second tier to the first tier. In exemplary embodiments of the invention, the first and second storage mediums are contained within one or more thin provisioning storage systems, and data is moved between the first and second tiers by allocating thin provisioning chunks to the data being moved. | 11-05-2009 |
20090327625 | MANAGING METADATA FOR DATA BLOCKS USED IN A DEDUPLICATION SYSTEM - Provided are a method, system, and article of manufacture for managing metadata for data blocks used in a deduplication system. File metadata is maintained for files having data blocks in a computer readable device. Data block metadata is maintained for each data block in the computer readable device. The data block metadata for one data block includes a data block reference and content identifier identifying content of the data block. The file metadata for each file includes the data block reference to each data block in the file. A determination is made of an unreferenced data block in the computer readable device that has become unreferenced. Indication is made that the data block metadata for the determined unreferenced data block as unreferenced metadata. The data block reference of the unreferenced metadata is maintained in the computer readable device in response to determining that a includes the data block indicated in the unreferenced metadata. | 12-31-2009 |
20100122049 | SYSTEM AND METHOD FOR MANAGING AN ELECTRONIC STORAGE VOLUME - A system and method for managing an electronic storage volume is described. The method includes assigning a threshold to a constrained storage space to define a first state in which an amount of data stored in the constrained storage space exceeds the threshold and a second state in which the amount of data stored in the confined storage space does not exceed the threshold. The method also includes comparing the amount of data to be stored in the constrained storage space and the threshold, and performing a predefined action if the comparison indicates that the amount data to be in the confined storage space would cause a transition between the first state and the second state. | 05-13-2010 |
20100169588 | OPTIMIZED MEMORY MANAGEMENT FOR RANDOM AND SEQUENTIAL DATA WRITING - A method and system writes data to a memory device including writing data to varying types of physical write blocks. The method includes receiving a request to write data for a logical block address within an LBA range to the memory device. Depending on whether the quantity of valid data in the memory device meets a predetermined criteria, the data is written to a specific chaotic block, a general chaotic block, or a mapped block. The mapped block is assigned for writing data for the LBA range, the specific chaotic block is assigned for writing data for contiguous LBA ranges including the LBA range, and the general chaotic block is assigned for writing data for any LBA range. Lower fragmentation and write amplification ratios may result by using this method and system. | 07-01-2010 |
20100191925 | DEFERRED VOLUME METADATA INVALIDATION - A method, system, and computer program product for managing modified metadata in a storage controller cache pursuant to a recovery action by a processor in communication with a memory device is provided. A count of modified metadata tracks for a storage rank is compared against a predetermined criterion. If the predetermined criterion is met, a storage volume having the storage rank is designated with a metadata invalidation flag to defer metadata invalidation of the modified metadata tracks until after the recovery action is performed. | 07-29-2010 |
20110107041 | METHOD FOR EXECUTING DATA UPDATES IN AN IC CARD - A method is for executing n data updates in an IC Card which has memory pages supporting m erase operations per page, with m05-05-2011 | |
20110113205 | DATA STORAGE DEVICE - A data storage device connectable to an electronic device and a remote storage serve, which comprises memory managing means configured to receive a request for a piece of information from the electronic device and if said piece of information is stored both in a local memory of the data storage device and in the remote server, retrieving the requested piece of information from the local memory to reduce latency. | 05-12-2011 |
20110276770 | METHOD AND SYSTEM FOR ANALYSING MOST RECENTLY USED REGISTRY KEYS - A method is disclosed for analysing a computing device including a non-volatile storage medium in which a set of snapshots is stored. Each snapshot comprises at least one most recently used, MRU, key for an application, at least one MRU key having a plurality of elements. The method comprises comparing a first MRU key from a first snapshot for a first time and a corresponding second MRU key for a second snapshot for a second time temporally following said first time. If the second MRU key has a second element identified as less recently used than a first element of the second MRU key and the second element of the first MRU key is identified as not being less recently used than the first element of the first MRU key, the first element is labelled as having been newly modified between the first and second times. | 11-10-2011 |
20120124305 | METHOD AND SYSTEM FOR INTEGRATING DATABASE MEMORY MANAGEMENT IN VIRTUAL MACHINES - Memory of a database management system (DBMS) that is running in a virtual machine is managed using techniques that integrate DBMS memory management with virtual machine memory management. Because of the integration, the effectiveness of DBMS memory management is preserved even though the physical memory allocated to the virtual machine may change during runtime as a result of varying memory demands of other applications, e.g., instances of other virtual machines, running on the same host computer as the virtual machine. | 05-17-2012 |
20120198187 | Technique for preserving memory affinity in a non-uniform memory access data processing system - Techniques for preserving memory affinity in a computer system is disclosed. In response to a request for memory access to a page within a memory affinity domain, a determination is made if the request is initiated by a processor associated with the memory affinity domain. If the request is not initiated by a processor associated with the memory affinity domain, a determination is made if there is a page ID match with an entry within a page migration tracking module associated with the memory affinity domain. If there is no page ID match, an entry is selected within the page migration tracking module to be updated with a new page ID and a new memory affinity ID. If there is a page ID match, then another determination is made whether or not there is a memory affinity ID match with the entry with the page ID field match. If there is no memory affinity ID match, the entry is updated with a new memory affinity ID; and if there is a memory affinity ID match, an access counter of the entry is incremented. | 08-02-2012 |
20120221812 | METHOD FOR PRESERVING MEMORY AFFINITY IN A NON-UNIFORM MEMORY ACCESS DATA PROCESSING SYSTEM - A method for preserving memory affinity in a computer system is disclosed. The method reduces and sometimes eliminates memory affinity loss due to process migration by restoring the proper memory affinity through dynamic page migration. The memory affinity access patterns of individual pages are tracked continuously. If a particular page is found almost always to be accessed from a particular remote access affinity domain for a certain number of times, and without any intervening requests from other access affinity domain, the page will migrate to that particular remote affinity domain so that the subsequent memory access becomes local memory access. As a result, the proper pages are migrated to increase memory affinity. | 08-30-2012 |
20120254564 | Managing High Speed Memory - A mechanism is provided for managing a high speed memory. An index entry indicates a storage unit in the high speed memory. A corresponding non-free index is set for a different type of low speed memory. The indicated storage unit in the high speed memory is assigned to a corresponding low speed memory by including the index entry in the non-free index. The storage unit in the high speed memory is recovered by demoting the index entry from the non-free index. The mechanism acquires a margin performance loss corresponding to a respective non-free index in response to receipt of a demotion request. The margin performance loss represents a change in a processor read operation time caused by performing a demotion operation in a corresponding non-free index. The mechanism compares the margin performance losses of the respective non-free indexes and selecting a non-free index whose margin performance loss satisfies a demotion condition as a demotion index. The mechanism selects an index entry from the demotion index to perform the demotion operation. | 10-04-2012 |
20120265954 | SYSTEM AND METHOD FOR OPTIMIZED RECLAMATION PROCESSING IN A VIRTUAL TAPE LIBRARY SYSTEM - A storage management application determines that a source virtual tape requires reclamation, identifies all block addresses for active data of a source virtual tape and sorts the block addresses in an ascending order, identifies a target virtual tape which has sufficient free capacity to store the active data of said source virtual tape and the last written block address on said target virtual tape, and sends a command to the VTL-system instructing it to perform reclamation including information about said source and said target virtual tape, the sorted list of block addresses denoting active data on the source virtual tape and the starting block address on the target virtual tape. The reclamation logic references the active data host blocks of said source volume to said target virtual tape starting at said starting block address by just updating the host block to disk block mapping table. | 10-18-2012 |
20130046943 | STORAGE CONTROL SYSTEM AND METHOD, AND REPLACING SYSTEM AND METHOD - A row buffer | 02-21-2013 |
20130086339 | METHOD AND APPARATUS FOR HIGH BANDWIDTH DICTIONARY COMPRESSION TECHNIQUE USING DELAYED DICTIONARY UPDATE - Method, apparatus, and systems employing novel delayed dictionary update schemes for dictionary-based high-bandwidth lossless compression. A pair of dictionaries having entries that are synchronized and encoded to support compression and decompression operations are implemented via logic at a compressor and decompressor. The compressor/decompressor logic operatives in a cooperative manner, including implementing the same dictionary update schemes, resulting in the data in the respective dictionaries being synchronized. The dictionaries are also configured with replaceable entries, and replacement policies are implemented based on matching bytes of data within sets of data being transferred over the link. Various schemes are disclosed for entry replacement, as well as a delayed dictionary update technique. The techniques support line-speed compression and decompression using parallel operations resulting in substantially no latency overhead. | 04-04-2013 |
20130111160 | SELECTIVE SPACE RECLAMATION OF DATA STORAGE MEMORY EMPLOYING HEAT AND RELOCATION METRICS | 05-02-2013 |
20130173875 | METHOD OF MANAGING STORAGE REGION OF MEMORY DEVICE, AND STORAGE APPARATUS USING THE METHOD - A method of managing a storage region of a memory device, and a storage apparatus using the method. In the method, data blocks are arranged in an order of performing writing on the memory device; a frequency of updating data in a logical page to be written is determined, based on whether an invalid physical page is present in a block within a window size that is initially set based on a most recently written data block from among the arranged data blocks, in response to a write request, and the data in the logical page is stored in a storage region of the memory device classified according to the determined frequency, according to the determined frequency. | 07-04-2013 |
20130205107 | METHOD AND ARRANGEMENT FOR RECORDING A MEDIA STREAM - The present invention relates to the field of digital media recording such as video or audio, where compression is used to reduce the amount of data to save onto a data storage. In particular the invention relates to recording media, wherein the memory area required to store the media is unknown. The invention discloses a method for recording digital streamed media with a number of media frames in a memory by encoding each media frame into an encoded frame comprising a first number of sets of data, representing different characteristics of the media stream. The invention further relates to detecting if the data storage is full and if the data storage is full, storing new frames in the data storage previously occupied by the data sets representing the least prioritized characteristics of the previously stored encoded frames. The invention also relates to a corresponding media recorder and computer program. | 08-08-2013 |
20130238866 | System and Method for Robust and Efficient Free Chain Management - Disclosed herein are systems, methods, and non-transitory computer-readable storage media for managing free chains of compute resources. A system configured to practice the method divides a free chain of compute resources into a usable part (UP) which contains resources available for immediate allocation and an unusable part (UUP) which contains resources not available for immediate allocation but which become available after a certain minimum number of allocations. The system sorts resources in the UP by block number, and maintains a last used object (LUO) vector, indexed by block number, which records a last object in the UP for each block. Each time the system frees a resource, the system adds the freed resource to a tail of the UUP and promotes an oldest resource in the UUP to the UP. This approach can manage free chains in a manner that is both flaw tolerant and has relatively high performance. | 09-12-2013 |
20130275693 | PREFERENTIAL BLOCK RECYCLING IN A REDIRECT-ON-WRITE FILESYSTEM - Provided are techniques for more efficient data storage on a computing system. An inode table is provided and populated with information relating to current and N previous locations within data storage that a particular data block has been stored. When a particular data block is modified in a redirect on write system, the modified data block is stored, if possible is a previous storage location for that particular data block and the current data location may be saved for use as the location for a subsequent modification. | 10-17-2013 |
20130297898 | Reducing Read Starvation - Apparatus, methods, and other embodiments associated with reducing read starvation that is supported by a multi-purpose buffer managed by a least recently used (LRU) data structure are described. One example method includes changing how certain retired pages are added back into the LRU. The dual-purpose buffer may be used in data de-duplication to support satisfying ingest requests and to support satisfying read requests. The method may also include controlling the LRU to allocate active ingest pages and active read pages from the head of the LRU. The method may also include controlling the LRU to cause the active ingest page and the active read page to re-enter the LRU at the tail of the LRU. Unlike conventional approaches, the method may also include controlling the LRU to cause a retired ingest page to re-enter the LRU at a location other than the tail (e.g., head) of the LRU. | 11-07-2013 |
20130326168 | MEMORY MANAGEMENT METHODS AND SYSTEMS FOR MOBILE DEVICES - Memory management methods and systems for mobile devices are provided. A memory usage of a memory is monitored by a built-in memory management component of an OS of the device and a user-oriented memory management component. It is determined whether the memory usage of the memory is greater than a first threshold or a second threshold, wherein the second threshold is greater than the first threshold. When the memory usage of the memory is greater than the first threshold, a multi-level memory management is performed by the user-oriented memory management component. When the memory usage of the memory is greater than the second threshold, a primitive memory management is performed by the built-in memory management component. | 12-05-2013 |
20140068207 | Reducing Page Faults in Host OS Following a Live Partition Mobility Event - Page faults during partition migration from a source computing system to a destination computing system are reduced by assigning each page used by a process as being hot or cold according to their frequency of use by the process. During a live partition migration, the cold or coldest (least frequently used) pages are copied to the destination server first, followed copying the warmer (less frequently used) and concluded by copying the hottest (most frequently used) pages. After all dirtied pages have been refreshed, cutover from the instance on the source server to the destination server is made. By transferring the warm and hot pages last (or later) in the migration process, the number of dirtied pages is reduced, thereby reducing page faults subsequent to the cutover. | 03-06-2014 |
20140089613 | MANAGEMENT OF DATA ELEMENTS OF SUBGROUPS - A plurality of subgroups with a least recently used (LRU) list of data elements associated with count variables. The LRU lists have a top entry to store a most recently used data element and a bottom entry to store a least recently used data element. If a data element is accessed, then increase the value of the count variable and move the accessed data element to the top entry of the LRU list of the subgroup associated with the data element. If the value of the count variable of the accessed data element of the top entry is greater than a value of a count variable of a data element of a bottom entry of a LRU list of a subgroup with a higher priority, then swap the data element of the bottom entry with the accessed data element of the top entry. | 03-27-2014 |
20140164721 | CLOUD MANAGEMENT OF DEVICE MEMORY BASED ON GEOGRAPHICAL LOCATION - A method for managing memory of a device is disclosed. A computer system collects information about use, by the device, of data in the memory of the device. The information collected by the computer system includes a time and a location for which each portion of the data is used by the device. The computer system identifies patterns of use, by the device, of each portion of the data based on the information collected. The computer system then selects one or more portions of the data that are not needed in the memory of the device based on the patterns of use by the device. | 06-12-2014 |
20140181433 | Storage Device and Method for Enabling Hidden Functionality - A storage device and method for enabling hidden functionality are provided. In one embodiment, a storage device is provided comprising an interface a memory, and a controller. The controller is configured to receive a series of read and/or write commands to the memory from the host device. If the series of read and/or write commands received from the host device matches an expected pattern of read and/or write commands, irrespective what data is being read or written by those commands, the controller enables a special functionality mode of the storage device. Other embodiments are possible, and each of the embodiments can be used alone or together in combination. | 06-26-2014 |
20140189268 | HIGH READ BLOCK CLUSTERING AT DEDUPLICATION LAYER - Methods, systems, and computer program products are provided for deduplicating data mapping a plurality of file blocks of selected data to a plurality of logical blocks, deduplicating the plurality of logical blocks to thereby associate each logical block with a corresponding physical block of a plurality of physical blocks located on a physical memory device, two or more of the corresponding physical blocks being non-contiguous with each other, determining whether one or more of the corresponding physical blocks are one or more frequently accessed physical blocks being accessed at a frequency above a threshold frequency and being referred to by a common set of applications, and relocating data stored at the one or more frequently accessed physical blocks to different ones of the plurality of physical blocks, the different ones of the plurality of physical blocks being physically contiguous. | 07-03-2014 |
20140215169 | APPARATUS AND METHOD FOR MANAGING DATA ACCESS COUNT - An apparatus counts, for each piece of data, an access count indicating a number of times of access to the each piece of data for every unit time so as to store a management information element including the access count and identification information identifying the each piece of data. The apparatus deletes, from the plurality of management information elements stored in a memory, a management information element that includes the access count having a value minimum among the plurality of management information elements, when a number of the plurality of management information elements reaches a predetermined number. The apparatus determines whether there is a piece of data that satisfies a condition related to rapid increase of access, based on the access counts included in the plurality of management information elements. | 07-31-2014 |
20140281299 | OPPORTUNISTIC MIGRATION OF MEMORY PAGES IN A UNIFIED VIRTUAL MEMORY SYSTEM - Techniques are disclosed for transitioning a memory page between memories in a virtual memory subsystem. A unified virtual memory (UVM) driver detects a page fault in response to a memory access request associated with a first memory page, where a local page table does not include an entry corresponding to a virtual memory address included in the memory access request. The UVM driver, in response to the page fault, executes a page fault sequence. The page fault sequence includes modifying the ownership state associated with the first memory page to be central-processing-unit-shared. The page fault sequence further includes scheduling the first memory page for migration from a system memory associated with a central processing unit (CPU) to a local memory associated with a parallel processing unit (PPU). One advantage of the disclosed approach is that the PPU accesses memory pages with greater efficiency. | 09-18-2014 |
20140297977 | MANAGING APPARATUS - A managing server includes a predicting unit and a controlling unit. The predicting unit predicts either a time or a time period at which frequency of accesses from a plurality of apparatuses to a storage device storing data therein will exceed a predetermined threshold value, on a basis of a status of the accesses from the plurality of apparatuses to the storage device. The controlling unit controls the storage device so as to transmit the data stored therein to at least one of the plurality of apparatuses, prior to the predicted time or time period. | 10-02-2014 |
20140351532 | MINIMIZING DESTAGING CONFLICTS - Destage grouping of tracks is restricted to a bottom portion of a least recently used (LRU) list without grouping the tracks at a most recently used end of the LRU list to avoid the destaging conflicts. The destage grouping of tracks is destaged from the bottom portion of the LRU list. | 11-27-2014 |
20140365738 | Systems and Methods for Memory Page Offloading in Multi-Processor Computer Systems - Systems and methods for memory page offloading in multi-processor computer systems. An example method may comprise: detecting, by a computer system, a memory pressure condition on a first node; invalidating a page table entry for a memory page residing on the first node; copying the memory page to a second node; and updating the page table entry for the memory page to reference the second node. | 12-11-2014 |
20150052319 | MEMORY MANAGEMENT METHODS AND SYSTEMS FOR PAGE-OUT MECHANISM - Memory management methods and systems for page-out mechanism are provided. A page-out mechanism is performed via an OS (Operating System) based on a parameter of the page-out mechanism, wherein the page-out mechanism moves data from a memory to a storage unit. A usage of a page-out partition in the storage unit is monitored. The parameter of the page-out mechanism is dynamically set according to the usage of the page-out partition, wherein when the usage is increased, the parameter of the page-out mechanism is decreased, and when the usage is decreased, the parameter of the page-out mechanism is increased. | 02-19-2015 |
20150089170 | METHOD AND APPARATUS FOR MANAGING MEMORY - An apparatus for managing a memory including a working region and a compression region is provided. The working region stores uncompressed data. The apparatus includes a management module and a compression/decompression module. According to a recent used index and a compression ratio of a set of target data stored in the working region, the management module determines whether to transfer the target data to the compression region. When the management module determines to transfer the target data to the compression region, the compression/decompression module compresses the target data and transfers the compressed target data to the compression region. | 03-26-2015 |
20150331808 | PACKET CAPTURE DEEP PACKET INSPECTION SENSOR - Systems and processes for filtering network traffic. In one example, an event based flow record aging process may be used to identify flow records to be removed from memory to provide space for newer flow records. In this example, a new incoming network packet may trigger a determination of the flow record to be removed from memory based on the hardware time stamps of the flow records stored in memory. Determining aging flow records in this way may advantageously reduce the amount of computer resources required to manage flow record aging when compared to traditional techniques that require keeping a timer and periodically performing a cleanup process to check the freshness date of each flow record. | 11-19-2015 |
20150363236 | DATA REUSE TRACKING AND MEMORY ALLOCATION MANAGEMENT - Exemplary methods, apparatuses, and systems receive a first request for a storage address at a first access time. Entries are added to first and second data structures. Each entry includes the storage address and the first access time. The first data structure is sorted in an order of storage addresses. The second data structure is sorted in an order of access times. A second request for the storage address is received at a second access time. The first access time is determined by looking up the entry in first data structure using the storage address received in the second request. The entry in the second data structure is looked up using the determined first access time. A number of entries in second data structure that were subsequent to the second entry is determined. A hit count for a reuse distance corresponding to the determined number of entries is incremented. | 12-17-2015 |
20160055097 | HETEROGENEOUS UNIFIED MEMORY - Inventive aspects include a heterogeneous unified memory section, which includes an extended unified memory space across a plurality of physical heterogeneous memory modules. A cold page reclamation logic section can receive and prioritize cold pages from a system memory. The cold pages can include a first subset of memory pages having a first type of memory data and a second subset of memory pages having a second type of memory data. For example, the cold pages can include anon-type memory pages and file-type memory pages. A dynamic tuning logic section can manage space allocation within the extended unified memory space. An intelligent page sort logic section can distribute the cold pages among different pools of physical heterogeneous memory modules based on varying characteristics of the pools, and based on the assigned priorities. | 02-25-2016 |
20160062805 | MANAGEMENT OF MEMORY PAGES - In a method for managing memory pages, responsive to determining that a server is experiencing memory pressure, one or more processors identifying a first memory page in a listing of memory pages in the server. The method further includes determining whether the first memory page corresponds to a logical partition (LPAR) of the server that is scheduled to undergo an operation to migrate data stored on memory pages of the LPAR to another server. The method further includes, responsive to determining that the first memory page does correspond to a LPAR of the server that is scheduled to undergo an operation to migrate data, determining whether to evict the first memory page based on a memory page state associated with the first memory page. The method further includes, responsive to determining to evict the first memory page, evicting data stored in the first memory page to a paging space. | 03-03-2016 |
20160196069 | MANAGEMENT OF MEMORY PAGES | 07-07-2016 |