Patent application number | Description | Published |
20080259092 | SYSTEMS AND METHODS FOR COLOR MANAGING CONTENT ELEMENTS ON A DISPLAY DEVICE - A system and method for color managing content elements on a display device are provided. The systems and methods for color managing content elements on a display device comprising using a set of color space characteristic values that describe a color space of the display device to create a profile having a multidimensional table encapsulating the display behavior as described by color space characteristic values for the display device. | 10-23-2008 |
20080304741 | Automatic detection of calibration charts in images - Methods and apparatuses for locating an embedded color chart in an image are described. In one exemplary method, an image that includes an embedded color chart is located without the intervention of the user. The embedded color chart is verified and used to create a color profile of the image. Furthermore, the orientation angle of the color chart is determined and the image orientation is fixed based on this angle. | 12-11-2008 |
20090303256 | DISPLAY-POINTER VISIBILITY - A method, apparatus, and system are provided for accentuating a pointer. An idle time period associated with a pointer relating to a display is monitored. A determination is made whether a termination of the idle time period associated with the pointer has occurred based upon the monitoring. A representation of the pointer is modified in response to the termination of the idle time period associated with the pointer. | 12-10-2009 |
20100271505 | System and Method for RAW Image Processing - An automated RAW image processing method and system are disclosed. A RAW image and metadata related to the RAW image are obtained from a digital camera or other source. The RAW image and the related metadata are automatically processed using an Operating System service of a processing device to produce a resulting image in an absolute color space. The resulting image is then made available to an application program executing on the processing device through an application program interface with the Operating System service. | 10-28-2010 |
20100309344 | CHROMA NOISE REDUCTION FOR CAMERAS - A system, apparatus, computer readable medium, and method for noise reduction in image capturing devices involving an edge-preserving blur window is disclosed. In one embodiment, the edge-preserving blur includes only those pixels in the blur window that are visually close to the blur window's current center pixel in its blurring calculation. Limiting the pixels considered in the blur to those that are visually close to the center pixel ensures that the image's colors are not blurred along color edges within the image. Light-product information taken from the image's metadata, for example, the camera sensor's gain level, may be used to adjust the blur filter parameters dynamically. This allows the method to perform the appropriate amount of processing depending on the lighting situation of the image that is currently being processed. | 12-09-2010 |
20100309345 | Radially-Based Chroma Noise Reduction for Cameras - A system, apparatus, computer readable medium, and method for radially-dependent noise reduction in image capturing devices involving an edge-preserving blur window are disclosed. In one embodiment, the edge-preserving blur includes only those pixels in the blur window that are within a threshold value of the blur window's current center pixel in its blurring calculation. By creating a threshold function that varies radially from the center of the image sensor's light intensity falloff function, a more appropriate threshold value can be chosen for each pixel, allowing for more noise farther from the center of the image, and allowing for less noise closer to the center of the image. Light-product information taken from the image's metadata may be used to scale the threshold value parameters dynamically. This allows the method to perform the appropriate amount of processing depending on the lighting situation of the image that is currently being processed. | 12-09-2010 |
20100309364 | CONTINUOUS AUTOFOCUS MECHANISMS FOR IMAGE CAPTURING DEVICES - At least certain embodiments described herein provide a continuous autofocus mechanism for an image capturing device. The continuous autofocus mechanism can perform an autofocus scan for a lens of the image capturing device and obtain focus scores associated with the autofocus scan. The continuous autofocus mechanism can determine an acceptable band of focus scores based on the obtained focus scores. Next, the continuous autofocus mechanism can determine whether a current focus score is within the acceptable band of focus scores. A refocus scan may be performed if the current focus score is outside of the acceptable band of focus scores. | 12-09-2010 |
20120036433 | Three Dimensional User Interface Effects on a Display by Using Properties of Motion - The techniques disclosed herein use a compass, MEMS accelerometer, GPS module, and MEMS gyrometer to infer a frame of reference for a hand-held device. This can provide a true Frenet frame, i.e., X- and Y-vectors for the display, and also a Z-vector that points perpendicularly to the display. In fact, with various inertial clues from accelerometer, gyrometer, and other instruments that report their states in real time, it is possible to track the Frenet frame of the device in real time to provide a continuous 3D frame-of-reference. Once this continuous frame of reference is known, the position of a user's eyes may either be inferred or calculated directly by using a device's front-facing camera. With the position of the user's eyes and a continuous 3D frame-of-reference for the display, more realistic virtual 3D depictions of the objects on the device's display may be created and interacted with by the user. | 02-09-2012 |
20140368725 | CONTINUOUS AUTOFOCUS MECHANISMS FOR IMAGE CAPTURING DEVICES - At least certain embodiments described herein provide a continuous autofocus mechanism for an image capturing device, The continuous autofocus mechanism can perform an autofocus scan for a lens of the image capturing device and obtain focus scores associated with the autofocus scan. The continuous autofocus mechanism can determine an acceptable band of focus scores based on the obtained focus scores. Next, the continuous autofocus mechanism can determine whether a current focus score is within the acceptable band of focus scores. A refocus scan may be performed if the current focus score is outside of the acceptable band of focus scores. | 12-18-2014 |
Patent application number | Description | Published |
20120076405 | AUTOMATIC DETECTION OF CALIBRATION CHARTS IN IMAGES - Methods and apparatuses for locating an embedded color chart in an image are described. In one exemplary method, an image that includes an embedded color chart is located without the intervention of the user. The embedded color chart is verified and used to create a color profile of the image. Furthermore, the orientation angle of the color chart is determined and the image orientation is fixed based on this angle. | 03-29-2012 |
20120176401 | Gesture Mapping for Image Filter Input Parameters - This disclosure pertains to systems, methods, and computer readable medium for mapping particular user interactions, e.g., gestures, to the input parameters of various image processing routines, e.g., image filters, in a way that provides a seamless, dynamic, and intuitive experience for both the user and the software developer. Such techniques may handle the processing of both “relative” gestures, i.e., those gestures having values dependent on how much an input to the device has changed relative to a previous value of the input, and “absolute” gestures, i.e., those gestures having values dependent only on the instant value of the input to the device. Additionally, inputs to the device beyond user-input gestures may be utilized as input parameters to one or more image processing routines. For example, the device's orientation, acceleration, and/or position in three-dimensional space may be used as inputs to particular image processing routines. | 07-12-2012 |
20120242852 | Gesture-Based Configuration of Image Processing Techniques - This disclosure pertains to apparatuses, methods, and computer readable medium for mapping particular user interactions, e.g., gestures, to the input parameters of various image filters, while simultaneously setting auto exposure, auto focus, auto white balance, and/or other image processing technique input parameters based on the appropriate underlying image sensor data in a way that provides a seamless, dynamic, and intuitive experience for both the user and the client application software developer. Such techniques may handle the processing of image filters applying location-based distortions as well as those image filters that do not apply location-based distortions to the captured image data. Additionally, techniques are provided for increasing the performance and efficiency of various image processing systems when employed in conjunction with image filters that do not require all of an image sensor's captured image data to produce their desired image filtering effects. | 09-27-2012 |
20130195353 | Digital Image Color Correction - Systems, methods, and computer readable media for performing color correction operations to address memory color artifacts in a manner suited for real-time operations. In general, techniques are disclosed for correcting memory color rendering artifacts in an image without performing color space conversions. In one implementations, hue-saturation-value (HSV) image correction values may be expressed solely in terms of an image's base red-green-blue (RGB) color space values. Once expressed in this manner, color correction may be applied to the image directly—without the need to convert the image's color space into and out of a working color space (e.g., an HSV color space). As no color space conversions are necessary, the disclosed techniques are well-suited to real-time operations. | 08-01-2013 |
20130201349 | DIGITAL CAMERA RAW IMAGE SUPPORT - RAW camera images may be processed by a computer system using either a particular application or a system level service. In either case, at least some parameters needed for the processing are preferably separated from the executable binary of the application or service, and are provided in separate, non-executable, data-only files. Each of these files can correspond to a particular camera or other imaging device. When a user of the system attempts to open a RAW image file from an unsupported device, the local system may contact a server for on-demand download and on-the-fly installation of the required support resource. | 08-08-2013 |
20130215141 | Using Render-Graph Characteristics to Optimize the Rendering of an Adjusted Image - A technique for optimizing the rendering of such complex render-graphs caches intermediate buffers of nodes that are expected to be re-used after they've been rendered. The render-graph is examined to determine the number of re-uses of each node's output buffer, and the buffer is cached in memory until all the re-uses of the buffer have occurred. Once all the re-uses of the buffer have occurred, the buffer is removed from the cache. This technique guarantees that for a given render-graph, no nodes will be re-rendered, resulting in improved render performance. | 08-22-2013 |
20130321677 | SYSTEMS AND METHODS FOR RAW IMAGE PROCESSING - Systems and methods for processing raw image data are provided. One example of such a system may include memory to store image data in raw format from a digital imaging device and an image signal processor to process the image data. The image signal processor may include data conversion logic and a raw image processing pipeline. The data conversion logic may convert the image data into a signed format to preserve negative noise from the digital imaging device. The raw image processing pipeline may at least partly process the image data in the signed format. The raw image processing pipeline may also include, among other things, black level compensation logic, fixed pattern noise reduction logic, temporal filtering logic, defective pixel correction logic, spatial noise filtering logic, lens shading correction logic, and highlight recovery logic. | 12-05-2013 |
20130321679 | SYSTEMS AND METHODS FOR HIGHLIGHT RECOVERY IN AN IMAGE SIGNAL PROCESSOR - Image sensors have finite ranges of illuminance that may be captured. When the sensors for particular pixels receive an amount of light exceeding these finite ranges, the pixel values clip to the maximum pixel value. Systems and methods for estimating pixel values that are clipped or near clipping are provided. In one example, a method for processing image data includes determining that a first channel of the image data is saturated or near saturation. The method further includes computing a highlight recovery value for the first channel based upon alternative channels in the image data that are not saturated or near saturation. The highlight recovery value is applied to the first channel. | 12-05-2013 |
20130328898 | Render Tree Caching - GPU fragment programs can be used to render images in a computer system. These fragment programs are generated from render trees, which specify one or more filters or functions to be applied to an input image to render an output image. It is not uncommon for successive frames to require application of substantially the same filters. Therefore, rather than regenerate and recompile new fragment programs for successive corresponding render trees, the render trees are substantially uniquely identified and cached. Thus, when a render tree is received, it can be identified, and this identifier (such as a hash) can be used to determine whether a corresponding fragment program has already been generated, compiled and cached. If so, the corresponding cached fragment program is retrieved and executed. If not, a fragment program for the newly received render tree is generated and cached. | 12-12-2013 |
20140240539 | GESTURE MAPPING FOR IMAGE FILTER INPUT PARAMETERS - This disclosure pertains to systems, methods, and computer readable medium for mapping particular user interactions, e.g., gestures, to the input parameters of various image processing routines, e.g., image filters, in a way that provides a seamless, dynamic, and intuitive experience for both the user and the software developer. Such techniques may handle the processing of both “relative” gestures, i.e., those gestures having values dependent on how much an input to the device has changed relative to a previous value of the input, and “absolute” gestures, i.e., those gestures having values dependent only on the instant value of the input to the device. Additionally, inputs to the device beyond user-input gestures may be utilized as input parameters to one or more image processing routines. For example, the device's orientation, acceleration, and/or position in three-dimensional space may be used as inputs to particular image processing routines. | 08-28-2014 |
20150093023 | Backwards Compatible Extended Image Format - Techniques are provided for encoding an extended image such that it is backwards compatible with existing decoding devices. An extended image format is defined such that the extended image format is consistent with an existing image format over the full range of the existing image format. Because the extended image format is consistent with the existing image format over the full range of the existing image format, additional image information that is included in an extended image can be extracted from the extended image. A base version of an image (expressed using the existing image format) may be encoded in a payload portion and the extracted additional information may be stored in a metadata portion of a widely supported image file format. | 04-02-2015 |
20150106768 | Three Dimensional User Interface Effects On A Display By Using Properties Of Motion - The techniques disclosed herein use a compass, MEMS accelerometer, GPS module, and MEMS gyrometer to infer a frame of reference for a hand-held device. This can provide a true Frenet frame, i.e., X- and Y-vectors for the display, and also a Z-vector that points perpendicularly to the display. In fact, with various inertial clues from accelerometer, gyrometer, and other instruments that report their states in real time, it is possible to track the Frenet frame of the device in real time to provide a continuous 3D frame-of-reference. Once this continuous frame of reference is known, the position of a user's eyes may either be inferred or calculated directly by using a device's front-facing camera. With the position of the user's eyes and a continuous 3D frame-of-reference for the display, more realistic virtual 3D depictions of the objects on the device's display may be created and interacted with by the user. | 04-16-2015 |