Patent application number | Description | Published |
20080250073 | SQL CHANGE TRACKING LAYER - Systems and methods that track changes in a database via a change tracking layer that enables separation of change tracking and change enumeration. Such an arrangement enables multiple change enumeration and sync technologies over a single change tracking layer, while reducing amount of tracking information that are maintained. | 10-09-2008 |
20080319957 | EXTENSIBLE COMMAND TREES FOR ENTITY DATA MODEL PLATFORM - Systems and methods that provide for a canonical representation in a structured form of a query, against a data model platform. A parser component can parse the SQL query to generate the abstract class that represent the query (command tree). Moreover, a view generation component can supply a mapping between a rich structure (e.g., on the client side) and the relational side. Accordingly, a query can be represented by an abstract class in form of a tree structure with nodes, which has metadata tied therewith. | 12-25-2008 |
20090037394 | EXTENSIBLE COMMAND EXECUTION FOR ENTITY DATA MODEL PLATFORM - Systems and methods that translate queries defined by an entity model, to a dialect recognizable by storage provider (e.g., relational store). A bridge component facilitates translation of a rich object structure into flat relational constructs, which can then be executed by a relational store. Accordingly, queries defined by applications in terms of entity concepts can readily be employed in conjunction with relational data stores. Similarly, results obtained from executing the query can be converted back to a form understandable by the application. | 02-05-2009 |
20090313436 | CACHE REGIONS - A cache region can be created in a cache in response to receiving a cache region creation request from an application. A storage request from the application can identify the cache region and one or more objects to be stored in the cache region. Those objects can be stored in the cache region in response to receiving the storage request. | 12-17-2009 |
20090313438 | DISTRIBUTED CACHE ARRANGEMENT - Systems and methods that aggregate memory capacity of multiple computers into a single unified cache, via a layering arrangement. Such layering arrangement is scalable to a plurality of machines and includes a data manager component, an object manager component and a distributed object manager component, which can be implemented in a modular fashion. Moreover, the layering arrangement can provide for an explicit cache tier (e.g., cache-aside architecture) that applications are aware about, wherein decision are made explicitly which objects to put/remove in such applications (as opposed to an implicit cache wherein application do not know the existence of the cache.) | 12-17-2009 |
20100228749 | OPEN TYPES FOR DISTRIBUTED SYSTEMS - Techniques for exchanging data are provided herein. An “open” data type is established on a computer system to define the structure of data. The open data type includes one or more predefined properties, and enables one or more extension properties to be optionally included. Data that is defined according to the open data type includes data values for the predefined properties, and may optionally include data values for extension properties. A second computer system may consider the extension properties to be extension properties or predefined properties of a data type known at the second computer system. In this manner, the type system that is used to exchange data between nodes need not necessarily be commonly known among the nodes, and/or may be dynamically modified, even for a particular data instance. Furthermore, queries may be performed against the data that query against the predefined properties and/or extension properties. | 09-09-2010 |
20100293332 | CACHE ENUMERATION AND INDEXING - In response to a request including a state object, which can indicate a state of an enumeration of a cache, the enumeration can be continued by using the state object to identify and send cache data. Also, an enumeration of cache units can be performed by traversing a data structure that includes object nodes, which correspond to cache units, and internal nodes. An enumeration state stack can indicate a current state of the enumeration, and can include state nodes that correspond to internal nodes in the data structure. Additionally, a cache index data structure can include a higher level table and a lower level table. The higher level table can have a leaf node pointing to the lower level table, and the lower level table can have a leaf node pointing to one of the cache units. Moreover, the lower level table can be associated with a tag. | 11-18-2010 |
20100293333 | MULTIPLE CACHE DIRECTORIES - A first portion of an identifier can be used to assign the identifier to a slot in a first directory. The identifier can identify a cache unit in a cache. It can be determined whether assignment of the identifier to the slot in the first directory will result in the identifier and one or more other identifiers being assigned to the same slot in the first directory. If so, then the technique can include (1) using a second portion of the identifier to assign the identifier to a slot in a second directory; and (2) assigning the one or more other identifiers to one or more slots in the second directory. In addition, it can be determined whether a directory in a cache lookup data structure includes more than one pointer. If not, then a parent pointer that points to the subject directory can be removed. | 11-18-2010 |
20100293338 | CACHE CLEANUP AND LATCHING - A low priority queue can be configured to list low priority removal candidates to be removed from a cache, with the low priority removal candidates being sorted in an order of priority for removal. A high priority queue can be configured to list high priority removal candidates to be removed from the cache. In response to receiving a request for one or more candidates for removal from the cache, one or more high priority removal candidates from the high priority queue can be returned if the high priority queue lists any high priority removal candidates. If no more high priority removal candidates remain in the high priority queue, then one or more low priority removal candidates from the low priority queue can be returned in the order of priority for removal. Write-only latches can also be used during write operations in a cache lookup data structure. | 11-18-2010 |
20130268914 | SERVICE LEVEL OBJECTIVE FOR CLOUD HOSTED APPLICATIONS - Embodiments are directed to declaratively managing software applications, dynamically applying configuration changes to a specified software application based on a service level objective (SLO) and to efficiently performing an impact analysis for a specified service level agreement SLO. | 10-10-2013 |
20140189639 | SERVICE LEVEL OBJECTIVE FOR CLOUD HOSTED APPLICATIONS - Embodiments are directed to declaratively managing software applications, dynamically applying configuration changes to a specified software application based on a service level objective (SLO) and to efficiently performing an impact analysis for a specified service level agreement SLO. | 07-03-2014 |