Patent application number | Description | Published |
20100271071 | Universal Inter-Layer Interconnect for Multi-Layer Semiconductor Stacks - A circuit arrangement and method utilize a universal, standardized inter-layer interconnect in a multi-layer semiconductor stack to facilitate interconnection and communication between functional units disposed on a stack of semiconductor dies. Each circuit layer in the multi-layer semiconductor stack is required to include an inter-layer interface region that is disposed at substantially the same topographic location such that when the semiconductor dies upon which such circuit layers are disposed are arranged together in a stack, electrical conductors disposed within each semiconductor die are aligned with one another to provide an inter-layer bus that is oriented vertically, or transversely, with respect to the individual circuit layers. Based upon a standardized placement of the inter-layer interface region in each circuit layer, and a standardized arrangement of electrical conductors associated with the inter-layer bus, each circuit layer may designed using a standardized template upon which the design features necessary to implement the inter-layer bus are already provided, thereby simplifying circuit layer design and the interconnection of functional units to the inter-layer bus. In addition, vertically-oriented supernodes may be defined within a semiconductor stack to provide multiple independently-operating nodes having functional units disposed in multiple circuit layers of the stack. | 10-28-2010 |
20120187570 | HYBRID BONDING TECHNIQUES FOR MULTI-LAYER SEMICONDUCTOR STACKS - A circuit arrangement and method utilize hybrid bonding techniques that combine wafer-wafer bonding processes with chip-chip and/or chip-wafer bonding processes to form a multi-layer semiconductor stack, e.g., by bonding together one or more sub-assemblies formed by wafer-wafer bonding together with other sub-assemblies and/or chips using chip-chip and/or chip-wafer bonding processes. By doing so, the advantages of wafer-wafer bonding techniques, such as higher interconnect densities, may be leveraged with the advantages of chip-chip and chip-wafer bonding techniques, such as mixing and matching chips with different sizes, aspect ratios, and functions. | 07-26-2012 |
20120198406 | UNIVERSAL INTER-LAYER INTERCONNECT FOR MULTI-LAYER SEMICONDUCTOR STACKS - An apparatus, program product and method facilitate the design of a multi-layer circuit arrangement incorporating a universal, standardized inter-layer interconnect in a multi-layer semiconductor stack to facilitate interconnection and communication between functional units disposed on a stack of semiconductor dies. Each circuit layer in the multi-layer semiconductor stack is required to include an inter-layer interface region that is disposed at substantially the same topographic location such that when the semiconductor dies upon which such circuit layers are disposed are arranged together in a stack, electrical conductors disposed within each semiconductor die are aligned with one another to provide an inter-layer bus that is oriented vertically, or transversely, with respect to the individual circuit layers. | 08-02-2012 |
20130009324 | UNIVERSAL INTER-LAYER INTERCONNECT FOR MULTI-LAYER SEMICONDUCTOR STACKS - An apparatus, program product and method facilitate the design of a multi-layer circuit arrangement incorporating a universal, standardized inter-layer interconnect in a multi-layer semiconductor stack to facilitate interconnection and communication between functional units disposed on a stack of semiconductor dies. Each circuit layer in the multi-layer semiconductor stack is required to include an inter-layer interface region that is disposed at substantially the same topographic location such that when the semiconductor dies upon which such circuit layers are disposed are arranged together in a stack, electrical conductors disposed within each semiconductor die are aligned with one another to provide an inter-layer bus that is oriented vertically, or transversely, with respect to the individual circuit layers. | 01-10-2013 |
20130011968 | HYBRID BONDING TECHNIQUES FOR MULTI-LAYER SEMICONDUCTOR STACKS - A circuit arrangement and method utilize hybrid bonding techniques that combine wafer-wafer bonding processes with chip-chip and/or chip-wafer bonding processes to form a multi-layer semiconductor stack, e.g., by bonding together one or more sub-assemblies formed by wafer-wafer bonding together with other sub-assemblies and/or chips using chip-chip and/or chip-wafer bonding processes. By doing so, the advantages of wafer-wafer bonding techniques, such as higher interconnect densities, may be leveraged with the advantages of chip-chip and chip-wafer bonding techniques, such as mixing and matching chips with different sizes, aspect ratios, and functions. | 01-10-2013 |
Patent application number | Description | Published |
20090171892 | Object Query Over Previous Query Results - A method, apparatus and program product is provided for querying an object database. In response to performing a first query, a first set of query results is generated from an object database. Metadata associated with and describing the first set of query results is generated. The first query results and the associated metadata are mapped to a results entity. A second set of query results is generated by querying the results entity using query terms from the metadata in response to performing a second query. | 07-02-2009 |
20090177680 | Generic Bijection With Graphs - A method, apparatus and program product is provided for accessing data in an abstract data store. In response to receiving a key value, an object is created by converting the key value into a key tuple. A value tuple is requested from the abstract data store corresponding to the key tuple. The key tuple and value tuple are converted to the object by a projector, which then stores the object on an object graph. An object is saved to the abstract data store by persisting the object onto the object graph. The object is detected on the object graph by the projector. In response to detecting the object, the object is converted to a key tuple and a value tuple by the projector. The object graph is then synchronized with the abstract data store by pushing the key and value tuples to the abstract data store. | 07-09-2009 |
20090265419 | Executing Applications at Servers With Low Energy Costs - Embodiments of the invention provide methods, systems, and articles of manufacture for managing and executing applications in a clustered server system. In one embodiment, an application may be installed at an application server having the associated lowest energy cost of maintenance, thereby lowering the cost of operating the system. In another embodiment, requests for services from the system may be routed to application servers having the lowest energy cost, thereby lowering the cost of operating the system. | 10-22-2009 |
20090265704 | Application Management for Reducing Energy Costs - Embodiments of the invention provide methods, systems, and articles of manufacture for managing and executing applications in a clustered server system. In one embodiment, an application may be installed at an application server having the associated lowest energy cost of maintenance, thereby lowering the cost of operating the system. In another embodiment, requests for services from the system may be routed to application servers having the lowest energy cost, thereby lowering the cost of operating the system. | 10-22-2009 |
20090276783 | Expansion and Contraction of Logical Partitions on Virtualized Hardware - A method, apparatus, and program product manage a plurality of resources of at least one logically partitioned computing system of the type that includes a plurality of logical partitions managed by a partition manager with an application level administrative console resident in a logical partition of the computing system. Each logical partition is allocated at least a portion of the plurality of resources. A user request to adjust the allocation of at least a portion of the resources using the administrative console is received. The resources of the logically partitioned computing system to adjust in order to satisfy the user request are determined using the application level administrative console. The application level administrative console accesses the partition manager through a resource allocation interface to adjust the determined resources of the logically partitioned computing system in order to satisfy the user request. | 11-05-2009 |
20120185500 | DATA STORAGE AND MANAGEMENT SYSTEM | 07-19-2012 |
20120303634 | In-Memory Data Grid Hash Scheme Optimization - Systems and methods of managing an in-memory data grid (IMDG) may involve conducting a data distribution analysis of the IMDG on a periodic basis, and selecting a hash scheme from a plurality of hash schemes based on the data distribution analysis. In one example, the selected hash scheme is used to conduct a repopulation of the IMDG, wherein the repopulation increases the distribution evenness of database records across the IMDG. | 11-29-2012 |
20120303675 | In-Memory Data Grid Hash Scheme Optimization - Systems and methods of managing an in-memory data grid (IMDG) may involve conducting a data distribution analysis of the IMDG on a periodic basis, and selecting a hash scheme from a plurality of hash schemes based on the data distribution analysis. In one example, the selected hash scheme is used to conduct a repopulation of the IMDG, wherein the repopulation increases the distribution evenness of database records across the IMDG. | 11-29-2012 |
20130246462 | AUTONOMIC CACHING FOR IN MEMORY DATA GRID QUERY PROCESSING - A method, system and computer program product for autonomic caching in an IMDG has been provided. A method for autonomic caching in an IMDG includes receiving from a client of the IMDG a request for a primary query in the IMDG. The method also includes associating the primary query with a previously requested sub-query related to the primary query. Finally, the method includes directing the sub-query concurrently with a directing of the primary query without waiting to receive a request for the sub-query from the client. In this way, the method can proactively predict a receipt of the request for a sub-query following a request for a primary query prior the actual receipt of the request for the sub-query. | 09-19-2013 |
20140052736 | CUSTOM OBJECT-IN-MEMORY FORMAT IN DATA GRID NETWORK APPLIANCE - Techniques are disclosed for implementing custom object-in-memory formats in a data grid network appliance. The techniques include maintaining a record of format definitions on a client device of the data grid and a corresponding record of format definitions on a server device of the data grid. Each format definition may indicate one or more attributes of an object class and data types and byte ranges of the attributes. The client device may serialize one or more objects for storage in the data grid based on respective format definitions associated with the one or more objects and retrieved from the record of format definitions maintained on the client device. Further, the server device may perform one or more data grid operations using format definitions retrieved from the record of format definitions maintained on the server device. | 02-20-2014 |
Patent application number | Description | Published |
20120278344 | PROXIMITY GRIDS FOR AN IN-MEMORY DATA GRID - Techniques are described for scaling an in-memory data grid to satisfy the needs of client applications requesting data from the in-memory data grid. The in-memory data grid may monitor a variety of factors, such as data traffic or number of received queries from a particular geographical region, to determine whether to create a proximity grid. A proximity grid is an extension of the in-memory data grid that services queries originating from a defined geographic region. Each proximity grid stores data objects relevant to the proximity grid's assigned geographic region. Queries originating from the client applications within the geographic regions are then routed to the corresponding proximity grid for execution. The proximity grid may be terminated when the need of the client applications within the geographic region has subsided. | 11-01-2012 |
20120311295 | SYSTEM AND METHOD OF OPTIMIZATION OF IN-MEMORY DATA GRID PLACEMENT - A method, system, and program product for optimizing distribution and availability of data partitions is disclosed. Placement logic is run for data partition distribution which optimizes a configuration of a memory data grid. A performance impact of adding new members to the memory data grid is provided. Replication of data partitions in the memory data grid is minimized. Optimum data partition distribution in the memory data grid is determined. | 12-06-2012 |
20130041971 | TECHNIQUE FOR IMPROVING REPLICATION PERSISTANCE IN A CACHING APPLICANCE STRUCTURE - A method for improving replication persistence in a caching appliance structure can begin when a primary catalog service receives a command to instantiate a data partition. The primary catalog service can manage a collective of caching appliances in a networked computing environment. The data partition can include a primary shard and at least one replica shard. The primary shard of the data partition can be stored within a memory space of a first caching appliance. The at least one replica shard of the data partition can be stored within a non-volatile storage space of a second caching appliance. The first and the second caching appliances can be separate physical devices. The memory space of the second caching appliance that could have been used to store the at least one replica shard can be available for storing primary shards for other data partitions, increasing the capacity of the collective. | 02-14-2013 |
20130042140 | TECHNIQUE FOR IMPROVING REPLICATION PERSISTANCE IN A CACHING APPLICANCE STRUCTURE - A method for improving replication persistence in a caching appliance structure can begin when a primary catalog service receives a command to instantiate a data partition. The primary catalog service can manage a collective of caching appliances in a networked computing environment. The data partition can include a primary shard and at least one replica shard. The primary shard of the data partition can be stored within a memory space of a first caching appliance. The at least one replica shard of the data partition can be stored within a non-volatile storage space of a second caching appliance. The first and the second caching appliances can be separate physical devices. The memory space of the second caching appliance that could have been used to store the at least one replica shard can be available for storing primary shards for other data partitions, increasing the capacity of the collective. | 02-14-2013 |
20130073809 | DYNAMICALLY ALTERING TIME TO LIVE VALUES IN A DATA CACHE - A TTL value for a data object stored in-memory in a data grid is dynamically adjusted. A stale data tolerance policy is set. Low toleration for staleness would mean that eviction is certain, no matter the cost, and high toleration would mean that the TTL value would be set based on total cost. Metrics to report a cost to re-create and re-store the data object are calculated, and the TTL value is adjusted based on calculated metrics. Further factors, such as, cleanup time to evict data from a storage site, may be considered in the total cost. | 03-21-2013 |
20130212340 | PARTITION AWARE QUALITY OF SERVICE FEATURE - A method for providing a partition aware quality of service feature may include receiving an indication of data to be stored in a distributed memory grid, determining a quality of service policy rule to be applied in relation to storage of the data in the memory grid based on the indication, and initiating storage of data blocks of the data in the memory grid. The data blocks may be provided with corresponding partition identifiers that facilitate retrieval of the data by indicating a location of storage of respective ones of the data blocks within the memory grid. The method may further include providing a quality of service token in association with the partition identifier based on the quality of service policy rule. | 08-15-2013 |
20130227232 | PARTITION AWARE QUALITY OF SERVICE FEATURE - A method for providing a partition aware quality of service feature may include receiving an indication of data to be stored in a distributed memory grid, determining a quality of service policy rule to be applied in relation to storage of the data in the memory grid based on the indication, and initiating storage of data blocks of the data in the memory grid. The data blocks may be provided with corresponding partition identifiers that facilitate retrieval of the data by indicating a location of storage of respective ones of the data blocks within the memory grid. The method may further include providing a quality of service token in association with the partition identifier based on the quality of service policy rule. | 08-29-2013 |
20140007052 | DYNAMICALLY CONFIGURING AN INTEGRATED DEVELOPMENT ENVIRONMENT | 01-02-2014 |
20140007053 | DYNAMICALLY CONFIGURING AN INTEGRATED DEVELOPMENT ENVIRONMENT | 01-02-2014 |
20140052744 | CUSTOM OBJECT-IN-MEMORY FORMAT IN DATA GRID NETWORK APPLIANCE - Techniques are disclosed for implementing custom object-in-memory formats in a data grid network appliance. The techniques include maintaining a record of format definitions on a client device of the data grid and a corresponding record of format definitions on a server device of the data grid. Each format definition may indicate one or more attributes of an object class and data types and byte ranges of the attributes. The client device may serialize one or more objects for storage in the data grid based on respective format definitions associated with the one or more objects and retrieved from the record of format definitions maintained on the client device. Further, the server device may perform one or more data grid operations using format definitions retrieved from the record of format definitions maintained on the server device. | 02-20-2014 |
20140074966 | INTELLIGENT MAPPING FOR AN ENTERPRISE GRID - A computer receives a request from a client application for data within an enterprise grid. The computer sends the data along with a program to monitor the application's use of the data. The computer subsequently receives from the program a subset of the data being utilized by the client application. Responsive to a subsequent request for the data, the computer sends to the client application only the data corresponding to the identified subset of data being utilized. | 03-13-2014 |
20140089260 | WORKLOAD TRANSITIONING IN AN IN-MEMORY DATA GRID - Embodiments of the present invention disclose a method, system, and computer program product for transitioning a workload of a grid client from a first grid server to a second grid server. A replication process is commenced transferring application state from the first grid server to the second grid server. Prior to completion of the replication process: the grid client is rerouted to communicate with the second grid server. The second grid server receives a request from the grid client. The second grid server determines whether one or more resources necessary to handle the request have been received from the first grid server. Responsive to determining that the one or more resources have not been received from the first grid server, the second grid server queries the first grid server for the one or more resources. The second grid server responds to the request from the grid client. | 03-27-2014 |
20140244953 | IDENTIFYING AND ACCESSING REFERENCE DATA IN AN IN-MEMORY DATA GRID - Embodiments relate to providing normalization techniques for reference data in an in-memory data grid. An aspect includes monitoring object creation and access in an in-memory data grid and identifying reference data in an object field of a plurality of object instances. A reference map for the object field is created and the reference map is replicated across all partitions of the in-memory data grid. The reference data of an embodiment is stored in the reference map and the object field is updated to identify the reference map. Accordingly, the reference data may be accessed using the created reference map. | 08-28-2014 |
20140244954 | IDENTIFYING AND ACCESSING REFERENCE DATA IN AN IN-MEMORY DATA GRID - Embodiments relate to providing normalization techniques for reference data in an in-memory data grid. An aspect includes monitoring object creation and access in an in-memory data grid and identifying reference data in an object field of a plurality of object instances. A reference map for the object field is created and the reference map is replicated across all partitions of the in-memory data grid. The reference data of an embodiment is stored in the reference map and the object field is updated to identify the reference map. Accordingly, the reference data may be accessed using the created reference map. | 08-28-2014 |
20150019584 | SELF-LEARNING JAVA DATABASE CONNECTIVITY (JDBC) DRIVER - Various embodiments include solutions for querying desired data from a database without pulling supererogatory data. In one embodiment, a method includes: obtaining an initial database access query between an application and the database at a self-learning JDBC driver; monitoring subsequent database access queries between the application and the database over a period; and generating a modified database access query for querying the database from the self-learning JDBC driver, the modified database access query based upon the subsequent database access queries between the application and the database over the period. | 01-15-2015 |
Patent application number | Description | Published |
20130036410 | Real-Time Temperature Sensitive Machine Level Code Compilation And Execution - Methods, apparatuses, and computer program products for real-time temperature sensitive machine level code compilation and execution are provided. Embodiments include compiling and executing, by a just-in-time (JIT) compiler, machine level code; during execution of the machine level code, determining, by the JIT compiler, an execution temperature for the machine level code, including reading temperature measurements from one or more processor temperature sensors; based on the determined execution temperature, identifying, by the JIT compiler, a portion of the machine level code that, when executed, caused temperature measurements of one or more processor temperature sensors to exceed a predetermined threshold temperature; recompiling, by the JIT compiler, the machine level code including modifying the identified portion to generate a new execution temperature that is lower than the previously determined execution temperature; and executing, by the JIT compiler, the recompiled machine level code. | 02-07-2013 |
20130046952 | Administering Thermal Distribution Among Memory Modules With Call Stack Frame Size Management - Administering thermal distribution among memory modules in a computing system that includes temperature sensors, where each temperature sensor measures temperature of a memory module and thermal distribution is effected by: determining, in real-time by a user-level application in dependence upon the temperature measurements of the temperature sensors, whether a memory module is overheated; if a memory module is overheated and if a current call stack frame is stored on the overheated memory module, increasing, by the user-level application, a size of the current call stack frame to fill remaining available memory space on the overheated memory module, ensuring a subsequent call stack frame is stored on a different memory module. | 02-21-2013 |
20130047142 | Thermal-Aware Source Code Compilation - Thermal-aware source code compilation including: receiving, by a compiler, an identification of a target computing system, the identification of the target computing system specifying temperature sensors that measure temperature of a memory module; compiling the source code into an executable application including inserting in the executable application computer program instructions for thermal-aware execution, the computer program instructions, when executed on the target computing system, carry out the steps of: retrieving temperature measurements of one or more of the target computing system's temperature sensors; determining, in real-time in dependence upon the temperature measurements, whether a memory module is overheated; if a memory module is overheated, entering a thermal-aware execution state including, for each memory allocation in the executable application, allocating memory on a different memory module than the overheated memory module; and upon the temperature sensors indicating the memory module is no longer overheated, exiting the thermal-aware execution state. | 02-21-2013 |
20130061007 | GENERATING CODE THAT CALLS FUNCTIONS BASED ON TYPES OF MEMORY - In an embodiment, in response to reading a declaration of a function that specifies a name of the function and a type of memory on which the function operates, the name of the function, a pointer to the function, and the type are saved to a template. In response to reading a call statement that specifies the name of the function and an identifier of an object, first code is generated. The first code, when executed, reads the pointer to the function from a virtual function table pointed to by the object, finds an entry in the virtual function table that represents the function, and reads the pointer from the entry in the virtual function table. The call statement, when executed, requests a call of the function. Second code is generated that, when executed, calls the function using the pointer read from the virtual function table. | 03-07-2013 |
20130067136 | Administering Thermal Distribution Among Memory Modules Of A Computing System - A computing system includes a number of memory modules and temperature sensors. Each temperature sensor measures a temperature of a memory module. In such a computing system a garbage collector during garbage collection, determines whether a temperature measurement of a temperature sensor indicates that a memory module is overheated and, if a temperature measurement of a temperature sensor indicates a memory module is overheated, the garbage collector reallocates one or more active memory regions on the overheated memory module to a non-overheated memory module. Reallocating the active memory regions includes copying contents of the active memory regions from the overheated memory module to the non-overheated memory module. | 03-14-2013 |
20130067155 | Memory Type-Specific Access Control Of A Field Of A Record - A computing system includes computer memory of a number of different memory types. An application program compiled for execution on the computing system controls access to a field of a record in the computer memory of the computing system by defining a record that includes one or more fields, the one or more fields including a restricted field having a specification of restricted accessibility when the restricted field is allocated in a particular memory type; allocating an instance of the record in memory of the particular memory type; and denying each attempted access of the restricted field while the record is allocated in the particular memory type. | 03-14-2013 |
20130103885 | ADMINISTERING THERMAL DISTRIBUTION AMONG MEMORY MODULES OF A COMPUTING SYSTEM - A computing system includes a number of memory modules and temperature sensors. Each temperature sensor measures a temperature of a memory module. In such a computing system a garbage collector during garbage collection, determines whether a temperature measurement of a temperature sensor indicates that a memory module is overheated and, if a temperature measurement of a temperature sensor indicates a memory module is overheated, the garbage collector reallocates one or more active memory regions on the overheated memory module to a non-overheated memory module. Reallocating the active memory regions includes copying contents of the active memory regions from the overheated memory module to the non-overheated memory module. | 04-25-2013 |
20130103919 | ADMINISTERING THERMAL DISTRIBUTION AMONG MEMORY MODULES WITH CALL STACK FRAME SIZE MANAGEMENT - Administering thermal distribution among memory modules in a computing system that includes temperature sensors, where each temperature sensor measures temperature of a memory module and thermal distribution is effected by: determining, in real-time by a user-level application in dependence upon the temperature measurements of the temperature sensors, whether a memory module is overheated; if a memory module is overheated and if a current call stack frame is stored on the overheated memory module, increasing, by the user-level application, a size of the current call stack frame to fill remaining available memory space on the overheated memory module, ensuring a subsequent call stack frame is stored on a different memory module. | 04-25-2013 |
20130104110 | THERMAL-AWARE SOURCE CODE COMPILATION - Thermal-aware source code compilation including: receiving, by a compiler, an identification of a target computing system, the identification of the target computing system specifying temperature sensors that measure temperature of a memory module; compiling the source code into an executable application including inserting in the executable application computer program instructions for thermal-aware execution, the computer program instructions, when executed on the target computing system, carry out the steps of: retrieving temperature measurements of one or more of the target computing system's temperature sensors; determining, in real-time in dependence upon the temperature measurements, whether a memory module is overheated; if a memory module is overheated, entering a thermal-aware execution state including, for each memory allocation in the executable application, allocating memory on a different memory module than the overheated memory module; and upon the temperature sensors indicating the memory module is no longer overheated, exiting the thermal-aware execution state. | 04-25-2013 |
20130104111 | REAL-TIME TEMPERATURE SENSITIVE MACHINE LEVEL CODE COMPILATION AND EXECUTION - Methods, apparatuses, and computer program products for real-time temperature sensitive machine level code compilation and execution are provided. Embodiments include compiling and executing, by a just-in-time (JIT) compiler, machine level code; during execution of the machine level code, determining, by the JIT compiler, an execution temperature for the machine level code, including reading temperature measurements from one or more processor temperature sensors; based on the determined execution temperature, identifying, by the JIT compiler, a portion of the machine level code that, when executed, caused temperature measurements of one or more processor temperature sensors to exceed a predetermined threshold temperature; recompiling, by the JIT compiler, the machine level code including modifying the identified portion to generate a new execution temperature that is lower than the previously determined execution temperature; and executing, by the JIT compiler, the recompiled machine level code. | 04-25-2013 |
20130110443 | GRANTING AUTHORITY IN RESPONSE TO DEFECT DETECTION | 05-02-2013 |
20130179731 | RECOVERING FROM A THREAD HANG - A method, computer-readable storage medium, and computer system are provided. In an embodiment, a detection is made that at least one thread, in a plurality of threads executing in a computer, has hung. A lock order of a plurality of locks used by the plurality of threads is determined. A determination is made that a first thread in the plurality of threads violates the lock order. After the determination that the first thread violates the lock order, the computer is restarted and the first thread that violates the lock order is scheduled on a reduced-speed processor. | 07-11-2013 |