Patent application number | Description | Published |
20090144228 | DATA PARALLEL PRODUCTION AND CONSUMPTION - A parallel execution of a query is performed with a plurality of producers using a partitioned data source, thereby producing output data. The output data is consumed with at least one consumer. The at least one consumer is monitored. A determination is made whether to stop producing output data by at least one of the producers prior to completion of the execution of the query based on the monitoring. | 06-04-2009 |
20090144232 | DATA PARALLEL SEARCHING - A query that includes a search operator and that identifies an input data source is received. The input data source is partitioned into a plurality of partitions. A parallel search through the partitions is performed for an element that could halt the search. The parallel search is performed using a plurality of parallel workers. One of the parallel workers generates a notification when the element is found by that worker. The notification notifies the other parallel workers that the search could be halted. Each of the parallel workers generates an output set based on results of the search. The output sets are merged into a merged output set. | 06-04-2009 |
20090144346 | PARTITIONING AND REPARTITIONING FOR DATA PARALLEL OPERATIONS - A query that identifies an input data source is rewritten to contain data parallel operations that include partitioning and merging. The input data source is partitioned into a plurality of initial partitions. A parallel repartitioning operation is performed on the initial partitions to generate a plurality of secondary partitions. A parallel execution of the query is performed using the secondary partitions to generate a plurality of output sets. The plurality of output sets are merged into a merged output set. | 06-04-2009 |
20090299958 | REORDERING OF DATA ELEMENTS IN A DATA PARALLEL SYSTEM - A query that identifies an input data source is received. The input data source is partitioned into a plurality of partitions. Each of the partitions includes a set of data elements with an associated set of indices for indicating an ordering of the data elements. A query type for a query operator in the received query is identified. It is determined whether a reordering of data elements will be performed based on the identified query type. The data elements in at least one of the partitions are reordered when it is determined based on the identified query type that reordering will be performed. | 12-03-2009 |
20090299959 | QUERY RESULT GENERATION BASED ON QUERY CATEGORY AND DATA SOURCE CATEGORY - A method includes receiving a query that identifies an input data source. A query category for a query operator in the received query is identified. A data source category for the input data source is also identified. A results object is generated based on the identified query category and the identified data source category. The results object supports at least one of random access and sequential access to results produced by the query operator. | 12-03-2009 |
20090300224 | COLLECTION WITH LOCAL LISTS FOR A MULTI-PROCESSOR SYSTEM - A method includes providing a collection that includes a plurality of local lists. Each local list is associated with a different processor or thread in a multi-processor system. An interface to the collection is provided for adding elements to and removing elements from the collection. An add operation is performed with a first processor in the multi-processor system using the interface to add a first element to the collection. The interface is configured to cause the first element to be added to the local list associated with the first processor. | 12-03-2009 |
20090300591 | COMPOSABLE AND CANCELABLE DATAFLOW CONTINUATION PASSING - Parallel tasks are created, and the tasks include a first task and a second task. Each task resolves a future. At least one of three possible continuations for each of the tasks is supplied. The three continuations include a success continuation, a cancellation continuation, and a failure continuation. A value is returned as the future of the first task upon a success continuation for the first task. The value from the first task is used in the second task to compute a second future. The cancellation continuation is supplied if the task is cancelled and the failure continuation is supplied if the task does not return a value and the task is not cancelled. | 12-03-2009 |
20090300630 | WAITING BASED ON A TASK GROUP - A method includes creating a first task group. A plurality of task object representations are added to the first task group. Each representation corresponds to one task object in a first plurality of task objects. A wait operation is performed on the first task group that waits for at least one of the task objects in the first plurality of task objects to complete. | 12-03-2009 |
20090327999 | IMMUTABLE TYPES IN IMPERITIVE LANGUAGE - An imperative framework having a more than one type is defined. At least one of the types is an immutable type, and the immutable type has at least one immutable field. | 12-31-2009 |
20100077384 | PARALLEL PROCESSING OF AN EXPRESSION - A method includes compiling an expression into executable code that is configured to create a data structure that represents the expression. The expression includes a plurality of sub-expressions. The code is executed to create the data structure. The data structure is evaluated using a plurality of concurrent threads, thereby processing the expression in a parallel manner. | 03-25-2010 |
20100146245 | PARALLEL EXECUTION OF A LOOP - A method of executing a loop over an integer index range of indices in a parallel manner includes assigning a plurality of index subsets of the integer index range to a corresponding plurality of threads, and defining for each index subset a start point of the index subset, an end point of the index subset, and a boundary point of the index subset positioned between the start point and the end point of the index subset. A portion of the index subset between the start point and the boundary point represents a private range and the portion of the index subset between the boundary point and the end point represents a public range. Loop code is executed by each thread based on the index subset of the integer index range assigned to the thread. | 06-10-2010 |
20100146522 | HANDLING EXCEPTIONS IN A DATA PARALLEL SYSTEM - A method of handling exceptions in a data parallel system includes forwarding exceptions thrown by concurrent worker tasks to a coordination task. The thrown exceptions are aggregated into an aggregation exception structure. It is determined whether the aggregation exception structure will be handled by an exception handler. The concurrent worker tasks are unwound when it is determined that the aggregation exception structure will be handled. | 06-10-2010 |
20100162211 | PROVIDING ACCESS TO A DATASET IN A TYPE-SAFE MANNER - A method of providing access to a dataset in a type-safe manner includes storing a dataset including a plurality of data elements and a corresponding plurality of order keys for indicating an ordering of the data elements. Each order key is associated with one of the data elements. An interface to the dataset is generated that is parameterized by an element type parameter and a key type parameter. The interface is configured to provide access to the data elements and the order keys in the dataset in a type-safe manner. | 06-24-2010 |
20100185840 | PROPAGATING UNOBSERVED EXCEPTIONS IN A PARALLEL SYSTEM - A method of handling an exception in a parallel system includes constructing a task object, executing a method with the task object, and catching an exception with the task object during execution of the method. The exception is propagated in response to the task object becoming inaccessible without the exception having been observed. | 07-22-2010 |
20100241808 | CACHE-LINE AWARE COLLECTION FOR RUNTIME ENVIRONMENTS - Target data is allocated into caches of a shared-memory multiprocessor system during a runtime environment. The target data includes a plurality of data items that are allocated onto separate cache lines. Each data item is allocated on a separate cache line regardless of the size of the cache line of the system. The data items become members of a wrapper types when data items are value types. The runtime environment maintains a set of wrapper types of various sizes that are of typical cache line sizes. Garbage data is inserted into the cache line in cases where data items are reference types and data is stored on a managed heap. The allocation also configures garbage collectors in the runtime environment not to slide multiple data items onto the same cache line. Other examples are included where a developer can augment the runtime environment to be aware of cache line sizes. | 09-23-2010 |
20100250564 | TRANSLATING A COMPREHENSION INTO CODE FOR EXECUTION ON A SINGLE INSTRUCTION, MULTIPLE DATA (SIMD) EXECUTION - A method of translating a comprehension into executable code for execution on a SIMD (Single Instruction, Multiple Data stream) execution unit, includes receiving a user specified comprehension. The comprehension is compiled into a first set of executable code. An intermediate representation is generated based on the first set of executable code. The intermediate representation is translated into a second set of executable code that is configured to be executed by a SIMD execution unit. | 09-30-2010 |
20100250613 | QUERY PROCESSING USING ARRAYS - A method of processing a query includes receiving a language integrated query including at least one operator, and operating on an input array with the at least one operator. An output array is generated by the at least one operator based on the operation on the input array. | 09-30-2010 |
20100274937 | PROVIDING LOCK-BASED ACCESS TO NODES IN A CONCURRENT LINKED LIST - A method of providing lock-based access to nodes in a concurrent linked list includes providing a plurality of striped lock objects. Each striped lock object is configured to lock at least one of the nodes in the concurrent linked list. An index is computed based on a value stored in a first node to be accessed in the concurrent linked list. A first one of the striped lock objects is identified based on the computed index. The first striped lock object is acquired, thereby locking and providing protected access to the first node. | 10-28-2010 |
20100333107 | LOCK-FREE BARRIER WITH DYNAMIC UPDATING OF PARTICIPANT COUNT - A method of executing an algorithm in a parallel manner using a plurality of concurrent threads includes generating a lock-free barrier that includes a variable that stores both a total participants count and a current participants count. The total participants count indicates a total number of threads in the plurality of concurrent threads that are participating in a current phase of the algorithm, and the current participants count indicates a total number of threads in the plurality of concurrent threads that have completed the current phase. The barrier blocks the threads that have completed the current phase. The total participants count is dynamically updated during execution of the current phase of the algorithm. The generating, blocking, and dynamically updating are performed by at least one processor. | 12-30-2010 |
20110179038 | PER-THREAD, PER-INSTANCE DATA STORAGE - A method of storing per-thread, per-instance data, includes identifying a unique index value corresponding to a first instance, identifying type parameters based on the identified index value, and instantiating a generic holder object based on the identified type parameters. The generic holder object includes a thread local field configured to store per-thread data that is local to the first instance. | 07-21-2011 |
20110231861 | CANCELLATION OF CONCURRENT WORKER OBJECTS - A method of performing a cancellation of concurrent worker objects includes creating a plurality of concurrent worker objects with a manager object. At least one cancellation object is generated with the manager object. The at least one cancellation object includes a cancel method and a register method that support concurrent calls. A callback function is registered by calling the register method with one of the concurrent worker objects. A cancellation of the concurrent worker objects is requested by calling the cancel method with the manager object. | 09-22-2011 |
20110239217 | PERFORMING A WAIT OPERATION TO WAIT FOR ONE OR MORE TASKS TO COMPLETE - A method of performing a wait operation includes creating a first plurality of tasks and a continuation task. The continuation task represents a second plurality of tasks. The continuation task and each of the tasks in the first plurality have an associated wait handle. The wait handles for the first plurality of tasks and the continuation task are stored in an array. A wait operation is performed on the array, thereby waiting for at least one of the tasks in the first and second pluralities to complete. | 09-29-2011 |
20120254139 | PROVIDING LOCK-BASED ACCESS TO NODES IN A CONCURRENT LINKED LIST - A method of providing lock-based access to nodes in a concurrent linked list includes providing a plurality of striped lock objects. Each striped lock object is configured to lock at least one of the nodes in the concurrent linked list. An index is computed based on a value stored in a first node to be accessed in the concurrent linked list. A first one of the striped lock objects is identified based on the computed index. The first striped lock object is acquired, thereby locking and providing protected access to the first node. | 10-04-2012 |
20130145136 | HANDLING EXCEPTIONS IN A DATA PARALLEL SYSTEM - A method of handling exceptions in a data parallel system includes forwarding exceptions thrown by concurrent worker tasks to a coordination task. The thrown exceptions are aggregated into an aggregation exception structure. It is determined whether the aggregation exception structure will be handled by an exception handler. The concurrent worker tasks are unwound when it is determined that the aggregation exception structure will be handled. | 06-06-2013 |