Patent application number | Description | Published |
20120303721 | HYBRID REQUEST/RESPONSE AND POLLING MESSAGING MODEL - A hybrid messaging model including a method that sends a first request message from a control process executing on a computer to a plurality of subordinate processes. The first request message directs the subordinate processes to enter a first state. An expected state is set equal to the first state in response to sending the first request message. A status message, including the expected state, is periodically broadcast from the control process to the plurality of subordinate processes. At least one confirmation message is received from each of the subordinate processes confirming that the subordinate process is in the expected state. Each of the confirmation messages is responsive to either the first request message or to the status message. A second request message is sent from the control process to the plurality of subordinate processes in response to receiving at least one confirmation message from each of the subordinate processes. | 11-29-2012 |
20120303913 | TRANSPARENT FILE SYSTEM MIGRATION TO A NEW PHYSICAL LOCATION - Transparent file system migration including a method that identifies a source physical storage location in a file system that supports simultaneous access by multiple independent processes. A target physical storage location in the file system is identified and its contents are migrated to the target physical storage location. The migrating includes disallowing new allocations to the source physical storage location and copying contents from the source physical storage location to the target physical storage location. A write request to write data to the source physical storage location is received prior to or during the migrating. The write request is serviced contemporaneously with the migrating. Read and write requests directed to the source physical storage location are redirected to the target physical storage location. The redirecting is responsive to the migrating completing. The requests are from the multiple independent processes and include both logical and physical input/outputs. | 11-29-2012 |
20130185661 | MONITORING AND DEBUGGING QUERY EXECUTION OBJECTS - Nodes in a query execution data structure have monitor and dump methods defined. The monitor method enables the collection of information from the node. The dump method outputs the monitored information from each node. A monitor and debug mechanism of the preferred embodiments includes a graphical user interface that allows a user to graphically examine a query execution tree, to enable monitoring of the nodes on a node-by-node basis, and to view information dumped from the query execution data structure as the query is executed or after the query is executed. The result is a powerful tool that allows efficiently monitoring and debugging a query implemented in an object oriented query execution data structure. | 07-18-2013 |
20130191411 | MONITORING AND DEBUGGING QUERY EXECUTION OBJECTS - Nodes in a query execution data structure have monitor and dump methods defined. The monitor method enables the collection of information from the node. The dump method outputs the monitored information from each node. A monitor and debug mechanism of the preferred embodiments includes a graphical user interface that allows a user to graphically examine a query execution tree, to enable monitoring of the nodes on a node-by-node basis, and to view information dumped from the query execution data structure as the query is executed or after the query is executed. The result is a powerful tool that allows efficiently monitoring and debugging a query implemented in an object oriented query execution data structure. | 07-25-2013 |
20140095508 | EFFICIENT SELECTION OF QUERIES MATCHING A RECORD USING A CACHE - A method is provided for constructing a cache for storing results of previously evaluated queries in a binary tree based on a cache key. The cache is searched, by a processing device, for a node representing a set of previously evaluated queries that match a given record using an instance of the cache key. In response to finding a representative node that matches the instance of the cache key, the method performs processing associated with the set of previously evaluated queries. In response to finding no representative nodes, the method creates a new node using the instance of the cache key. The method of the invention then updates the cache by classifying the compatibility of newly registered queries with the cache key and evaluating whether compatible newly registered queries match the record. | 04-03-2014 |
20140095531 | DYNAMIC OUTPUT SELECTION USING HIGHLY OPTIMIZED DATA STRUCTURES - A node container is constructed with multiple binary tree views. A processing device creates a primary view binary tree within the node container based on a primary key. A set of records that match fields of the primary key is chained to the primary view. The processing device also creates one or more alternate views within the node container to allow access to the same set of records via one or more alternate keys, which are subsets of the fields of the primary key. The chain of records is extracted from the node container when matched to a subsequent query search. | 04-03-2014 |
20140101164 | EFFICIENT SELECTION OF QUERIES MATCHING A RECORD USING A CACHE - A method is provided for constructing a cache for storing results of previously evaluated queries in a binary tree based on a cache key. The cache is searched, by a processing device, for a node representing a set of previously evaluated queries that match a given record using an instance of the cache key. In response to finding a representative node that matches the instance of the cache key, the method performs processing associated with the set of previously evaluated queries. In response to finding no representative nodes, the method creates a new node using the instance of the cache key. The method of the invention then updates the cache by classifying the compatibility of newly registered queries with the cache key and evaluating whether compatible newly registered queries match the record. | 04-10-2014 |
Patent application number | Description | Published |
20090150560 | REAL-TIME MINING AND REDUCTION OF STREAMED DATA - A stream data node receives real-time streamed data from one or more input devices, dynamically filters the streamed data to reduce the streamed data, and delivers the reduced data when requested. By providing real-time filtering of the data, the amount of data that must be stored in a database may be substantially reduced. The stream data node can perform aggregation functions, group functions, and select functions, thereby also significantly reducing the amount of data that must be stored in a database. The stream data node may also be part of a query execution data structure, where it delivers its data when requested by another node in the query execution data structure. | 06-11-2009 |
20090248618 | Database Query Processing with Dynamically Activated Buffers - A method, apparatus and program product dynamically activate and/or deactivate buffers during execution of a database query. The dynamic deactivation and activation is based on the monitoring of information associated with the consumption of data by a plurality of threads during execution of an access plan for the database query. By doing so, the selection of a less optimal buffer position during optimization may be replaced with a more optimal buffer position during runtime, often resulting in improved query performance and parallelism. | 10-01-2009 |
20100036804 | Maintained and Reusable I/O Value Caches - Embodiments of the invention provide techniques for maintaining I/O value caches for database queries. Each maintained cache may be configured for use with a particular database query. Each cache may be persistently maintained in a system, meaning the cache is not automatically deleted after some period of time, and may thus be used to process subsequent instances of the same query. By use of the maintained cache, executing subsequent instances of the query may be avoided, thus saving time and system resources. Further, the maintained cache may be adapted to process other queries having similar characteristics to the initial query. The data included in each cache may be refreshed as required by changes to the underlying data. | 02-11-2010 |
20100036805 | System Maintainable and Reusable I/O Value Caches - Embodiments of the invention provide techniques for maintaining I/O value caches for database queries. Each maintained cache may be configured for use with a particular database query. Each cache may be persistently maintained in a system, meaning the cache is not automatically deleted after some period of time, and may thus be used to process subsequent instances of the same query. By use of the maintained cache, executing subsequent instances of the query may be avoided, thus saving time and system resources. Further, the maintained cache may be adapted to process other queries having similar characteristics to the initial query. The data included in each cache may be refreshed as required by changes to the underlying data. | 02-11-2010 |