Patent application number | Description | Published |
20100201871 | Temporally consistent caption detection on videos using a 3D spatiotemporal method - A caption detection system wherein all detected caption boxes over time for one caption area are identical, thereby reducing temporal instability and inconsistency. This is achieved by grouping candidate pixels in the 3D spatiotemporal space and generating a 3D bounding box for one caption area. 2D bounding boxes are obtained by slicing the 3D bounding boxes, thereby reducing temporal instability as all 2D bounding boxes corresponding to a caption area are sliced from one 3D bounding box and are therefore identical over time. | 08-12-2010 |
20100208994 | Filling holes in depth maps - Various implementations relate to improving depth maps. This may be done, for example, by identifying bad depth values and modifying those values. The values may represent, for example, holes and/or noise. According to a general aspect, a segmentation is determined based on an intensity image. The intensity image is associated with a corresponding depth image that includes depth values for corresponding locations in the intensity image. The segmentation is applied to the depth image to segment the depth image into multiple regions. A depth value is modified in the depth image based on the segmentation. A two-stage iterative procedure may be used to improve the segmentation and then modify bad depth values in the improved segmentation, and iterating until a desired level of smoothness is achieved. Both stages may be based, for example, on average depth values in a segment. | 08-19-2010 |
20100220932 | SYSTEM AND METHOD FOR STEREO MATCHING OF IMAGES - A system and method for stereo matching of at least two images, e.g., a stereoscopic image pair, employing a global optimization function, e.g., a belief propagation function, that utilizes dynamic programming as a preprocessing step are provided. The system and method of the present disclosure provide for acquiring a first image and a second image from a scene, estimating the disparity of at least one point in the first image with at least one corresponding point in the second image, and minimizing the estimated disparity using a belief propagation function, e.g., a global optimization function, wherein the belief propagation function is initialized with a result of a deterministic matching function, e.g., dynamic programming, applied to the first and second image to speed up the belief propagation function. The system and method further generates a disparity map from the estimated disparity and converts the disparity map into a depth map. | 09-02-2010 |
20100315410 | SYSTEM AND METHOD FOR RECOVERING THREE-DIMENSIONAL PARTICLE SYSTEMS FROM TWO-DIMENSIONAL IMAGES - A system and method for recovering three-dimensional (3D) particle systems from two-dimensional (2D) images are provided. The system and method of the present invention provide for identifying a fuzzy object in a two-dimensional (2D) image; selecting a particle system from a plurality of predetermined particle systems, the selected particle system relating to a predefined fuzzy object; generating at least one particle of the selected particle system; simulating the at least one particle to update states of the at least one particle; rendering the selected particle system; comparing the rendered particle system to the identified fuzzy object in the 2D image; and storing the selected particle system if the comparison result is within an acceptable threshold, wherein the stored particle system represents the recovered geometry of the fuzzy object. | 12-16-2010 |
20110044531 | SYSTEM AND METHOD FOR DEPTH MAP EXTRACTION USING REGION-BASED FILTERING - A system and method for extracting depth information from at least two images employing region-based filtering for reducing artifacts are provided. The present disclosure provides a post-processing algorithm or function for reducing the artifacts generated by scanline Dynamic Programming (DP) or other similar methods. The system and method provides for acquiring a first image and a second image from a scene, estimating the disparity of at least one point in the first image with at least one corresponding point in the second image to generate a disparity map, segmenting at least one of the first or second images into at least one region, and filtering the disparity map based on the segmented regions. Furthermore, anisotropic filters are employed, which have a great smoothing effect along the vertical direction than that of the horizontal direction, and therefore, reduce stripe artifacts without significantly blurring the depth boundaries. | 02-24-2011 |
20110069064 | SYSTEM AND METHOD FOR DEPTH EXTRACTION OF IMAGES WITH FORWARD AND BACKWARD DEPTH PREDICTION - A system and method for spatiotemporal depth extraction of images with forward and backward depth prediction are provided. The system and method of the present disclosure provide for acquiring a plurality of frames, generating a first depth map of a current frame in the plurality of frames based on a depth map of a previous frame in the plurality of frames, generating a second depth map of the current frame in the plurality of frames based on a depth map of a subsequent frame in the plurality of frames, and processing the first depth map and the second depth map to produce a third depth map for the current frame. | 03-24-2011 |
20110085727 | SYSTEM AND METHOD FOR MARKING A STEREOSCOPIC FILM - A system and method for marking a stereoscopic film with colors are provided. The system and method provides for marking a left image with a mark and a right image with a mark having complementary colors, wherein upon viewing, the marks are not visible under certain conditions. The system and method provide for acquiring a stereoscopic image, the stereoscopic image including a first image and a second image, applying a first mark to the first image in a predetermined location, the first mark having a first color, and applying a second mark to the second image in substantially the same predetermined location as in the first image, the second mark having a second color that is different than the first color of the first mark, wherein when viewed in three-dimensional mode, the first mark and the second mark combine into a single mark of one color. | 04-14-2011 |
20110110583 | SYSTEM AND METHOD FOR DEPTH EXTRACTION OF IMAGES WITH MOTION COMPENSATION - A system and method for spatiotemporal depth extraction of images are provided. The system and method provide for acquiring a sequence of images from a scene, the sequence including a plurality of successive frames of images, estimating the disparity of at least one point in a first image with at least one corresponding point in a second image for at least one frame, estimating motion of the at least one point in the first image, estimating the disparity of the at least one next successive frame based on the estimated disparity of at least one previous frame in a forward direction of the sequence, wherein the estimate disparity is compensated with the estimated motion, and minimizing the estimated disparity of each of the plurality of successive frames based on the estimated disparity of at least one previous frame in a backward direction of the sequence. | 05-12-2011 |
20110142309 | SYSTEM AND METHOD FOR MEASURING POTENTIAL EYESTRAIN OF STEREOSCOPIC MOTION PICTURES - A system and method for measuring the potential eyestrain felt by audiences while watching a 3D presentation, e.g., a stereoscopic motion picture, are provided. The eyestrain measurement system and method of the present disclosure is based on the measurement of disparity (or depth) and disparity transition of stereoscopic images of the 3D presentation. The system and method of the present disclosure provide for acquiring a first image and a second image from a first segment, estimating disparity of at least one point in the first image with at least one corresponding point in the second image, estimating disparity transition of a sequence of first and second images, and determining potential eyestrain felt while viewing the 3D presentation based on the disparity and the disparity transition of the sequence of the first and second images. | 06-16-2011 |
20110184807 | System and Method for Filtering Targeted Advertisements for Video Content Delivery - In accordance with an embodiment, a method of inserting advertisements into video content includes electronically filtering a first list of advertisements according to user preference data to determine a second list of advertisements. The video content has a plurality of segments, each segment of which is associated with a category from the plurality of categories. Furthermore, each advertisement in the first list of advertisements is associated with a video category from a plurality of categories, and electronically filtering includes filtering the first list of advertisements for the plurality of video segments on a segment by segment basis. The method further includes transmitting the second list of advertisements to a user device for insertion with the video content. | 07-28-2011 |
20110185381 | System and Method for Matching Targeted Advertisements for Video Content Delivery - In accordance with an embodiment, a method of matching video content to advertising content includes electronically receiving a video content metadata from a content provider, and matching the video content metadata to advertising content metadata of a global list of advertisements. The video content metadata corresponds to video content being sent to a user device and includes at least one keyword. Furthermore, the advertising content metadata corresponds to advertising content and includes at least one keyword. Matching the video content metadata to advertising content metadata includes comparing the at least one keyword of the video content metadata to the at least one keyword of the advertising content metadata. | 07-28-2011 |
20110185384 | System and Method for Targeted Advertisements for Video Content Delivery - In accordance with an embodiment, a method of inserting advertisements into video content includes electronically receiving video content metadata from a content provider, matching the video content metadata to advertising content metadata to provide a first list of advertisements, filtering the first list of advertisements according to user preference data to determine a second list of advertisements, and transmitting the second list of advertisements to a user device for insertion into the video content. The video content metadata corresponds to video content being sent to a user device, and the advertising content metadata corresponds to advertising content. | 07-28-2011 |
20110293022 | MESSAGE PASSING INTERFACE (MPI) FRAMEWORK FOR INCREASING EXECUTION SPEEDAULT DETECTION USING EMBEDDED WATERMARKS - A system and method for processing video uses a message protocol to communicate between computing units. An image request message is sent to an administrator process of a master node from at least one slave process to request an image to process. Responsive to the request message, an image name message is sent to a requesting slave process from the administrator process to retrieve the image from a queue. The image associated with the image name is processed. Images to process are requested until a completion message is received from the administrator process. | 12-01-2011 |
20120114184 | TRAJECTORY-BASED METHOD TO DETECT AND ENHANCE A MOVING OBJECT IN A VIDEO SEQUENCE - The present invention concerns a method and associated apparatus for using a trajectory-based technique to detect a moving object in a video sequence, such as the ball in a soccer game. In one embodiment, the method comprises steps of identifying and evaluating sets of connected components in a video frame, filtering the list of connected components by comparing features of the connected components to predetermined criteria, identifying candidate trajectories across multiple frames, evaluating the candidate trajectories to determine a selected trajectory, and processing images in the video sequence based at least in part upon the selected trajectory. | 05-10-2012 |
20120121174 | METHOD FOR DETECTING AND ADAPTING VIDEO PROCESSING FOR FAR-VIEW SCENES IN SPORTS VIDEO - A method is disclosed for analyzing video to detect far-view scenes in sports video to determine when certain image processing algorithms should be applied. The method comprises analyzing and classifying the fields of view of images from a video signal, creating and classifying the fields of view of sets of sequential images, and selectively applying image processing algorithms to sets of sequential images representing a particular type of field of view. | 05-17-2012 |
20120224629 | OBJECT-AWARE VIDEO ENCODING STRATEGIES - A method of object-aware video coding is provided that comprises the steps of: receiving a video sequence having a plurality of frames; selecting at least two frames; determing total area of at least one object of interest in each of the at least two frames; comparing the total area to a threshold area; classifying each of the at least two frames as being a low object weighted frame or a high object weighted frame, low object weighted frames being frames having the total area exceeding the threshold area and high object weighted frames being frame having the total area not exceeding the threshold area; and encoding each low object weighted frame according to one encoding mode and encoding each high object weighted frame according to a different encoding mode. | 09-06-2012 |
20120263437 | HUMAN INTERACTION TRAJECTORY-BASED SYSTEM - A method and associated apparatus for using a trajectory-based technique to detect a moving object in a video sequence at incorporates human interaction through a user interface. The method comprises steps of identifying and evaluating sets of connected components in a video frame, filtering the list of connected components by comparing features of the connected components to predetermined criteria, identifying candidate trajectories across multiple frames, evaluating the candidate trajectories to determine a selected trajectory, eliminating incorrect trajectories through use of the interface and processing images in said video sequence responsive to the evaluating and eliminating steps. | 10-18-2012 |
20120281139 | System and Method for Video Caption Re-Overlaying for Video Adaptation and Retargeting - In accordance with an embodiment, a method of processing an electronic image having caption text includes receiving the electronic source image, detecting the caption text in the electronic source image, reformatting the electronic source image, reformatting the caption text, and overlaying the reformatted caption text on the reformatted electronic image to form a resultant image. | 11-08-2012 |
20120288015 | DATA PRUNING FOR VIDEO COMPRESSION USING EXAMPLE-BASED SUPER-RESOLUTION - Methods and apparatuses for data pruning for video compression using example-based super resolution are provided. A method and apparatus for encoding is provided in which patches of video are extracted from input video, grouped together using a clustering method, and representative patches are packed into patch frames. The original video is downsized and sent either along with, or in addition to, the patch frames. At a decoder, the method and apparatus provided extract patches from the patch frames and create a patch library. The regular video frames are upsized and the low resolution patches are replaced by patches from the patch library by searching the library using the patches in the decoded regular frames as keywords. If there are no appropriate patches, no replacement is made. A post processing procedure is used to enhance the spatiotemporal smoothness of the recovered video. | 11-15-2012 |
20120288217 | HIGH DYNAMIC RANGE (HDR) IMAGE SYNTHESIS WITH USER INPUT - A new high dynamic range image synthesis which can handle the local object motion, wherein an interactive graphical user interface is provided for the end user, through which one can specify the source image for separate part of the final high dynamic range image, either by creating a image mask or scribble on the image. The high dynamic range image synthesis includes the following steps: capturing low dynamic range images with different exposures; registering the low dynamic range images; estimating camera response function; converting the low dynamic range images to temporary radiance images using estimated camera response function; and fusing the temporary radiance images into a single high dynamic range (HDR) image by employing a method of layered masking. | 11-15-2012 |
20120294369 | METHODS AND APPARATUS FOR SAMPLING-BASED SUPER RESOLUTION VIDEO ENCODING AND DECODING - Methods and apparatus are provided for sampling-based super resolution video encoding and decoding. The encoding method receives high resolution pictures and generates low resolution pictures and metadata there from, the metadata for guiding post-decoding post-processing of the low resolution pictures and the metadata; and then encodes the low resolution pictures and the metadata using at least one encoder. The corresponding decoding method receives a bitstream and decodes low resolution pictures and metadata there from using a decoder; and then reconstructs high resolution pictures respectively corresponding to the low resolution pictures using the low resolution pictures and the metadata. | 11-22-2012 |
20130028330 | Methods and Apparatus for Reducing Vector Quantization Error Through Patch Shifting - Methods and apparatus are provided for reducing vector quantization error through patch shifting. A method generates, from an input video sequence, one of more high resolution replacement patches, the one or more high resolution replacement patches for replacing one or more low resolution patches during a reconstruction of the input video sequence. This generating step generates the one or more high resolution replacement patches using data corresponding to a patch spatial shifting process, the patch spatial shifting process for reducing jittery artifacts caused by a motion-induced vector quantization error in the one or more high resolution replacement patches, the data for at least deriving a patch size of the one or more high resolution replacement patches such that the one or more high resolution replacement patches are generated to have the patch size greater than a patch size of the one or more low resolution patches in order to be suitable for use in the patch spatial shifting process. | 01-31-2013 |
20130061231 | CONFIGURABLE COMPUTING ARCHITECTURE - A configurable computing system for parallel processing of software applications includes an environment abstraction layer (EAL) for abstracting low-level functions to the software applications; a space layer including a distributed data structure; and a kernel layer including a job scheduler for executing parallel processing programs constructing the software applications according to a configurable mode. | 03-07-2013 |
20130089107 | Method and Apparatus for Multimedia Queue Management - Methods and systems for a multimedia queue management solution that maintaining graceful Quality of Experience (QoE) degradation are provided. The method selects a frame from all weighted queues based on a gradient function indicating a network performance rate change and a distortion rate caused by the frame and its related frames in the queue, and dropping the selected frame and all its related frames, and continues to drop similarly chosen frame until a network performance rate change caused by the dropping frame and its related frames meets a predetermined performance metric. A frame gradient is a distortion rate divided by a network performance rate change caused by the frame and its related frames, and a distortion rate is based on a sum of each individual frame distortion rate when the frame and its related frames are replaced by some other frames derived from remaining frames based on a replacement method. | 04-11-2013 |
20130177085 | Systems and Methods for Video Denoising - A method for video denoising, the method includes extracting a current pixel from a current video frame and n-1 pixels from respective n-1 previous video frames to form an n-point pixel data set, wherein the current pixel has an initial value; temporally filtering the current pixel using the n-point pixel data set to generate a temporal estimated true value for the pixel; estimating a quantity of outliers in the n-point pixel data set; if the quantity of outliers is greater than or equal to a threshold, spatially filtering the current pixel using neighboring pixels in the current video frame to generate an spatial estimated true value for the pixel, and using the spatial estimated true value for the pixel as a final estimated value for the pixel; if the quantity of outliers is less than the threshold, using the temporal estimated true value for the pixel as the final estimated value for the pixel. | 07-11-2013 |
20130251284 | System and Method for Filtering an Image - An embodiment is configured to calculate a perceptual masking factor at a pixel location at a block boundary of the image, calculate a parameter for a filter at the pixel location at the block boundary, and filter the image around the pixel location at the block boundary employing the filter with the calculated parameter. The perceptual masking factor is formed as a product of a background activity masking map and a luminance masking map. The filter includes a parameter that is selected in view of a quality of experience performance measure for the image at the pixel location at the block boundary of the image. | 09-26-2013 |
20130286288 | System and Method for Encoder-Integrated Media Denoising - Embodiments are provided herein to achieve video or image sequence encoding with an improved denoising algorithm that is both efficient computationally and has acceptable overhead cost in comparison to other denoising schemes for video encoding. The embodiments include using recursive bilateral filtering as part of the denoising algorithm, which is integrated into a video encoder to overcome limitations of other encoder-integrated denoising algorithms. An embodiment method includes receiving, at a filtering and residual computation function at the encoder, a macro block comprising a plurality of pixels. The filtering and residual computation function also receives, from a motion estimation function at the encoder, a reference block. The reference block comprises a plurality of reference pixels corresponding to the macro block. The filtering and residual computation function further applies a recursive bilateral filter function to each of the pixels of the macro block using the reference pixels of the reference block. | 10-31-2013 |
20130293725 | No-Reference Video/Image Quality Measurement with Compressed Domain Features - Techniques for objectively determining perceived video/image quality, the techniques including receiving a degraded bit-stream comprising encoded video/image data, and subsequently parsing the bit-stream to extract one or more video/image coding components. The video coding components may include intra-prediction modes, discrete cosine transform (DCT) coefficients, motion information, or combinations thereof, and may be used as a basis for objectively predicting a Quality of Experience (QoE) or Motion Opinion Score (MOS) score of the degraded bit-stream. | 11-07-2013 |
20140169663 | System and Method for Video Detection and Tracking - System and method embodiments are provided to enable features and functionalities for automatically detecting and localizing the position of an object in a video frame and tracking the moving object in the video over time. One method includes detecting a plurality of objects in a video frame using a combined Histograms of Oriented Gradients (HOG) and Local Binary Pattern (LBP) algorithm, highlighting the detected objects, and tracking one of the detected objects that is selected by a user in a plurality of subsequent video frames. Also included is a user device configured to detect a plurality of objects in a video frame displayed on a display screen coupled to the user device using a combined HOG and LBP algorithm, highlight the detected objects, and track one of the detected objects that is selected by a user in a plurality of subsequent video frames on the display screen. | 06-19-2014 |