Patent application number | Description | Published |
20080317051 | Methods and System for Highly Ordered Transaction Processing - Improved techniques are disclosed for processing transactions or requests in a data processing system. For example, a method for processing requests in a system including a plurality of nodes includes the following steps. At least two nodes of the plurality of nodes receive a plurality of requests. The two nodes exchange information to determine an order for processing requests. The two nodes process the requests in accordance with the order. The order may include a total order or a partial order. | 12-25-2008 |
20090037910 | METHODS AND SYSTEMS FOR COORDINATED TRANSACTIONS IN DISTRIBUTED AND PARALLEL ENVIRONMENTS - Automated techniques are disclosed for coordinating request or transaction processing in a data processing system. For example, a technique for handling compound requests, in a system comprising multiple nodes for executing requests in which an individual request is associated with a particular node, comprises the following steps. A compound request comprising at least two individual requests associated with a same node is received. It is determined if both of the at least two individual requests are executable. The compound request is executed if it is determined that all individual requests of the compound request can execute. | 02-05-2009 |
20090037913 | METHODS AND SYSTEMS FOR COORDINATED TRANSACTIONS - Automated techniques are disclosed for coordinating request or transaction processing in a data processing system. For example, a technique for handling requests in a data processing system comprises the following steps. A compound request comprising at least two individual requests of different types is received. An individual request r | 02-05-2009 |
20100011123 | Method and Apparatus for Hierarchical Selective Personalization - A technique for delivering content in a client-server system based on a request received at a computing device from a client includes determining a current load on a next-level computing device of a hierarchy. When this current load is such that a response time for delivery of the request from the next-level computing device would increase above a given threshold, a client type associated with the request is checked and, when the client type indicates that the client is below a given priority level, content to be delivered to the client in response to the request is personalized at the receiving computing device. When the current load is such that the response time would not increase above a given threshold, the request is sent from the receiving computing device to the next-level computing device and the content to be delivered is personalized at the next-level computing device. | 01-14-2010 |
20110153373 | TWO-LAYER DATA ARCHITECTURE FOR RESERVATION MANAGEMENT SYSTEMS - A reservation management system includes at least one processing node that includes a memory and a processor. The at least one processing node further includes a set of reservation processing and transaction modules that manage and process reservation requests and inquiries. At least one general purpose database is communicatively coupled to the at least one processing node. The at least one general purpose database includes a set of pre-allocated tables of fixed length records. At least one persistent storage device is communicatively coupled to the at least one processing node. The general purpose database persistently stores the set of pre-allocated tables of fixed length records on the at least one persistent storage device. | 06-23-2011 |
20110258629 | METHODS AND SYSTEMS FOR COORDINATED TRANSACTIONS IN DISTRIBUTED AND PARALLEL ENVIRONMENTS - Automated techniques are disclosed for minimizing communication between nodes in a system comprising multiple nodes for executing requests in which a request type is associated with a particular node. For example, a technique comprises the following steps. Information is maintained about frequencies of compound requests received and individual requests comprising the compound requests. For a plurality of request types which frequently occur in a compound request, the plurality of request types is associated to a same node. | 10-20-2011 |
20120078963 | SUPPORTING LINKED MULTI-USER DECISION MAKING IN ENVIRONMENTS WITH CONSTRAINED SHARED RESOURCES UTILIZING DURABLE FILES - Embodiments of the present invention manage multiple requests to allocate real world resources in a multi-user environment. A set of resource availability information is stored in a first durable data file for each resource in a plurality of resources provided by a database environment. The database environment is shared between a plurality of users. A decision context is associated with a second durable data file. The decision context is associated with a user interacting with the database environment. The decision context exists for a defined duration of time. A least one resource is determined to have been temporarily allocated to the decision context for the defined duration of time. The second durable data file is updated to indicate that the at least one resource has been temporarily allocated to the decision context. The first durable data file is updated to indicate that the at least one resource is currently unavailable. | 03-29-2012 |
20120079212 | ARCHITECTURE FOR SHARING CACHES AMONG MULTIPLE PROCESSES - Various embodiments of the present invention provide a system for caching information in a multi-process environment. The system includes a processor. A shared memory is communicatively coupled to the processor. The shared memory includes a set of data. A writer process is communicatively coupled to the shared memory. The write process reads and updates the set of data. A plurality of reader processes is communicatively coupled to the shared memory. Each reader process reads at least part of the set of data directly from the shared memory and sends a set of update information to the writer process. The writer process then updates the set of data stored in the shared memory based on the set of update information. | 03-29-2012 |
20120079213 | MANAGING CONCURRENT ACCESSES TO A CACHE - Various embodiments of the present invention manage concurrent accesses to a resource in a parallel computing environment. A plurality of locks is assigned to manage concurrent access to a plurality of parts of a resource. A usage of at least one of the plurality of parts of the resource is monitored. The assignment of the plurality of locks to the plurality of parts of the resource is modified based on the usage that has been monitored. | 03-29-2012 |
20120079391 | SUPPORTING LINKED MULTI-USER DECISION MAKING IN ENVIRONMENTS WITH CONSTRAINED SHARED RESOURCES - Embodiments of the present invention manage multiple requests to allocate real world resources in a multi-user environment. A request for interacting with a database environment comprising records of allocations of a plurality of resources is received from a user in a plurality of users. The database environment is shared between the plurality of users. A set of action choices available for the request is provided to the user via the user interface. A set of resources required by each action choice is identified. The set of resources is associated with a decision context. The decision context exists for a period of time. The set of resources are allocated to the user for a duration of the decision context. The allocating prevents the set of resources from being allocated to other users for the duration of the decision context irrespective of a set of actions performed by the other users. | 03-29-2012 |
20120303908 | MANAGING CONCURRENT ACCESSES TO A CACHE - Various embodiments of the present invention allow concurrent accesses to a cache. A request to update an object stored in a cache is received. A first data structure comprising a new value for the object is created in response to receiving the request. A cache pointer is atomically modified to point to the first data structure. A second data structure comprising an old value for the cached object is maintained until a process, which holds a pointer to the old value of the cached object, at least one of one of ends and indicates that the old value is no longer needed. | 11-29-2012 |
20130282985 | MANAGING CONCURRENT ACCESSES TO A CACHE - Various embodiments of the present invention allow concurrent accesses to a cache. A request to update an object stored in a cache is received. A first data structure comprising a new value for the object is created in response to receiving the request. A cache pointer is atomically modified to point to the first data structure. A second data structure comprising an old value for the cached object is maintained until a process, which holds a pointer to the old value of the cached object, at least one of one of ends and indicates that the old value is no longer needed. | 10-24-2013 |
20140122320 | METHODS AND SYSTEMS FOR COORDINATED TRANSACTIONS IN DISTRIBUTED AND PARALLEL ENVIRONMENTS - Automated techniques are disclosed for minimizing communication between nodes in a system comprising multiple nodes for executing requests in which a request type is associated with a particular node. For example, a technique comprises the following steps. Information is maintained about frequencies of compound requests received and individual requests comprising the compound requests. For a plurality of request types which frequently occur in a compound request, the plurality of request types is associated to a same node. As another example, a technique for minimizing communication between nodes, in a system comprising multiple nodes for executing a plurality of applications, comprises the steps of maintaining information about an amount of communication between said applications, and using said information to place said applications on said nodes to minimize communication among said nodes. | 05-01-2014 |
20140123155 | METHODS AND SYSTEMS FOR COORDINATED TRANSACTIONS IN DISTRIBUTED AND PARALLEL ENVIRONMENTS - Automated techniques are disclosed for minimizing communication between nodes in a system comprising multiple nodes for executing requests in which a request type is associated with a particular node. For example, a technique comprises the following steps. Information is maintained about frequencies of compound requests received and individual requests comprising the compound requests. For a plurality of request types which frequently occur in a compound request, the plurality of request types is associated to a same node. As another example, a technique for minimizing communication between nodes, in a system comprising multiple nodes for executing a plurality of applications, comprises the steps of maintaining information about an amount of communication between said applications, and using said information to place said applications on said nodes to minimize communication among said nodes. | 05-01-2014 |
20140215159 | MANAGING CONCURRENT ACCESSES TO A CACHE - Various embodiments of the present invention allow concurrent accesses to a cache. A request to update an object stored in a cache is received. A first data structure comprising a new value for the object is created in response to receiving the request. A cache pointer is atomically modified to point to the first data structure. A second data structure comprising an old value for the cached object is maintained until a process, which holds a pointer to the old value of the cached object, at least one of one of ends and indicates that the old value is no longer needed. | 07-31-2014 |