Patent application number | Description | Published |
20100046635 | TILING IN VIDEO DECODING AND ENCODING - Implementations are provided that relate, for example, to view tiling in video encoding and decoding. A particular method includes accessing a video picture that includes multiple pictures combined into a single picture, accessing information indicating how the multiple pictures in the accessed video picture are combined, decoding the video picture to provide a decoded representation of at least one of the multiple pictures, and providing the accessed information and the decoded video picture as output. Some other implementations format or process the information that indicates how multiple pictures included in a single video picture are combined into the single video picture, and format or process an encoded representation of the combined multiple pictures. | 02-25-2010 |
20100135391 | METHODS AND APPARATUS FOR MOTION SKIP MOVE WITH MULTIPLE INTER-VIEW REFERENCE PICTURES - There are provided methods and apparatus for motion skip mode with multiple inter-view reference pictures. An apparatus includes an encoder for encoding an image block relating to multi-view video content by performing a selection, for the image block, of at least one of an inter-view reference picture list from a set of inter-view reference picture lists, an inter-view reference picture from among a set of inter-view reference pictures, and a disparity vector from among a set of disparity vectors corresponding to the inter-view reference picture. The encoder extracts motion information for the image block based on at least one of the inter-view reference picture list, the inter-view reference picture, and disparity vector. | 06-03-2010 |
20100284466 | VIDEO AND DEPTH CODING - Various implementations are described. Several implementations relate to video and depth coding. One method includes selecting a component of video information for a picture. A motion vector is determined for the selected video information or for depth information for the picture. The selected video information is coded based on the determined motion vector. The depth information is coded based on the determined motion vector. An indicator is generated that the selected video information and the depth information are coded based on the determined motion vector. One or more data structures are generated that collectively include the coded video information, the coded depth information, and the generated indicator. | 11-11-2010 |
20110001792 | VIRTUAL REFERENCE VIEW - Various implementations are described. Several implementations relate to a virtual reference view. According to one aspect, coded information is accessed for a first-view image. A reference image is accessed that depicts the first-view image from a virtual-view location different from the first-view. The reference image is based on a synthesized image for a location that is between the first-view and the second-view. Coded information is accessed for a second-view image coded based on the reference image. The second-view image is decoded. According to another aspect, a first-view image is accessed. A virtual image is synthesized based on the first-view image, for a virtual-view location different from the first-view. A second-view image is encoded using a reference image based on the virtual image. The second-view is different from the virtual-view location. The encoding produces an encoded second-view image. | 01-06-2011 |
20110038418 | CODE OF DEPTH SIGNAL - Various implementations are described. Several implementations relate to determining, providing, or using a depth value representative of an entire coding partition. According to a general aspect, a first portion of an image is encoded using a first-portion motion vector that is associated with the first portion and is not associated with other portions of the image. The first portion has a first size. A first-portion depth value is determined that provides depth information for the entire first portion and not for other portions. A second portion of an image is encoded using a second-portion motion vector that is associated with the second portion and is not associated with other portions of the image. The second portion has a second size that is different from the first size. A second-portion depth value is determined that provides depth information for the entire second portion and not for other portions. | 02-17-2011 |
20110142138 | REFINED DEPTH MAP - Various implementations are described. Several implementations relate to a refined depth map. According to one aspect, depth information for a picture in a set of pictures is accessed. Modified depth information for the picture is accessed. A refinement is determined that characterizes a difference between the depth information and the modified depth information. The refinement, and the depth information, is provided for use in processing one or more pictures in the set of pictures. | 06-16-2011 |
20110148858 | VIEW SYNTHESIS WITH HEURISTIC VIEW MERGING - Several implementations relate to view synthesis with heuristic view merging for 3D Video (3DV) applications. According to one aspect, a first candidate pixel from a first warped reference view and a second candidate pixel from a second warped reference view are assessed based on at least one of a backward synthesis process to assess a quality of the first and second candidate pixels, a hole distribution around the first and second candidate pixels, or on an amount of energy around the first and second candidate pixels above a specified frequency. The assessing occurs as part of merging at least the first and second warped reference views into a signal synthesized view. Based on the assessing, a result is determined for a given target pixel in the single synthesized view. The result may be determining a value for the given target pixel, or marking the given target pixel as a hole. | 06-23-2011 |
20110157229 | VIEW SYNTHESIS WITH HEURISTIC VIEW BLENDING - Various implementations are described. Several implementations relate to view synthesis with heuristic view blending for 3D Video (3DV) applications. According to one aspect, at least one reference picture, or a portion thereof, is warped from at least one reference view location to a virtual view location to produce at least one warped reference. A first candidate pixel and a second candidate pixel are identified in the at least one warped reference. The first candidate pixel and the second candidate pixel are candidates for a target pixel location in a virtual picture from the virtual view location. A value for a pixel at the target pixel location is determined based on values of the first and second candidate pixels. | 06-30-2011 |
20110268177 | JOINT DEPTH ESTIMATION - Various implementations are described. Several implementations relate to joint depth estimation for multiple depth maps. In one implementation, a first-view depth indicator for a location in a first view is estimated, and a second-view depth indicator for a corresponding location in a second view is estimated. The estimating of one or more of the first-view depth indicator and the second-view depth indicator is based on a constraint. The constraint provides a relationship between the first-view depth indicator and the second-view depth indicator for corresponding locations. | 11-03-2011 |
20110273529 | CODING OF DEPTH MAPS - Various implementations are described. Several implementations relate to filtering of depth maps. According to a general aspect, a first depth picture is accessed that corresponds to a first video picture. For a given portion of the first depth picture, a co-located video portion of the first video picture is determined. A video motion vector is accessed that indicates motion of the co-located video portion of the first video picture with respect to a second video picture. A second depth picture is accessed that corresponds to the second video picture. A depth portion of the second depth picture is determined, from the given portion of the first depth picture, based on the video motion vector. The given portion of the first depth picture is updated based on the depth portion of the second depth picture. | 11-10-2011 |
20110286530 | Frame packing for video coding - Implementations are provided that relate, for example, to view tiling in video encoding and decoding. A particular implementation accesses a video picture that includes multiple pictures combined into a single picture, and accesses additional information indicating how the multiple pictures in the accessed video picture are combined. The accessed information includes spatial interleaving information and sampling information. Another implementation encodes a video picture that includes multiple pictures combined into a single picture, and generates information indicating how the multiple pictures in the accessed video picture are combined. The generated information includes spatial interleaving information and sampling information. A bitstream is formed that includes the encoded video picture and the generated information. Another implementation provides a data structure for transmitting the generated information. | 11-24-2011 |
20110292043 | Depth Map Coding to Reduce Rendered Distortion - Several implementations relate to depth map coding. In one implementation, a depth coding rate, that results from coding one or more portions of a depth map using a particular coding mode, is determined. The depth map can be used to render video for a different view than that of the depth map. A depth map distortion, that results from coding the one or more portions of the depth map using the particular coding mode, is determined. A value of distortion for the rendered video, based on the depth map distortion and on a particular relationship between the depth map distortion and values of distortion for the rendered video, is determined. It is determined whether to use the particular coding mode to code the one or more portions of the depth map, and the determination is based on the value of distortion for the rendered video and the depth coding rate. | 12-01-2011 |
20110292044 | DEPTH MAP CODING USING VIDEO INFORMATION - Several implementations relate to depth map coding. In one implementation, it is determined that differences between collocated video blocks are small enough to be interchanged. Based on that determination, a depth block corresponding to a first of the video blocks is coded using an indicator that instructs a decoder to use a collocated depth block, corresponding to a second of the video blocks, in place of the depth block. In another implementation, a video signal includes a coding of at least a single indicator that instructs a decoder to decode both a depth block and a corresponding video block using collocated blocks, from other pictures, in place of the depth block and the corresponding video block. In another implementation, the depth block and the corresponding video block are decoded, based on the single indicator, using the collocated blocks in place of the depth block and the corresponding video block. | 12-01-2011 |
20110298895 | 3D VIDEO FORMATS - Several implementations relate to 3D video formats. One or more implementations provide adaptations to MVC and SVC to allow 3D video formats to be used. According to a general aspect, a set of images including video and depth is encoded. The set of images is related according to a particular 3D video format, and are encoded in a manner that exploits redundancy between the set of images. The encoded images are arranged in a bitstream in a particular order, based on the particular 3D video format that relates to the images. The particular order is indicated in the bitstream using signaling information. According to another general aspect, a bitstream is accessed that includes the encoded set of images. The signaling information is also accessed. The set of images is decoded using the signaling information. | 12-08-2011 |
20120044322 | 3D VIDEO CODING FORMATS - Several implementations relate to 3D video (3DV) coding formats. One implementation encodes multiple pictures that describe different three-dimensional (3D) information for a given view at a given time. Syntax elements are generated that indicate, for the encoded multiple pictures, how the encoded picture fits into a structure that supports 3D processing. The structure defines content types for the multiple pictures. A bitstream is generated that includes the encoded multiple pictures and the syntax elements. The inclusion of the syntax elements provides, at a coded-bitstream level, indications of relationships between the encoded multiple pictures in the structure. The syntax elements also enable efficient inter-layer coding of the 3DV content, thereby reducing the bandwidth used to transmit the 3DV content. Corresponding decoding implementations are also provided. Extraction methods are also provided from extracting pictures of interest from such a 3DV multiple pictures and the syntax elements, the video stream characterized by such a 3D structure. | 02-23-2012 |
20120056981 | INTER-LAYER DEPENDENCY INFORMATION FOR 3DV - Various implementations are directed to providing inter-layer dependency information. In one implementation, syntax elements are generated that indicate an inter-layer dependency structure among three-dimensional video (3DV) layers. Based on the inter-layer dependency structure, an inter-layer reference is identified for a picture from a layer of the 3DV layers. The picture is encoded based, at least in part, on the inter-layer reference. Corresponding decoding implementations are also provided. Additionally, in another implementation, a transmission priority and an indication of network congestion are used to determine whether to transmit data for a particular 3DV layer. The transmission priority is based on an inter-layer dependency structure among multiple 3DV layers. Another implementation is directed to a network abstraction layer unit that can explicitly identify and convey inter-layer references and corresponding dependencies. | 03-08-2012 |
20120140819 | DEPTH MAP CODING - Various implementations relate to depth map coding. In one method, a depth coding rate and depth distortion are determined for a coding mode. Based on the value of depth distortion, a correlation coefficient is determined between at least a portion of a video picture and a translated version of the video picture. The video picture is one or more of a video picture corresponding to the depth being coded, or a rendered video picture for a different view. A video distortion is determined based on the correlation coefficient, and is used to evaluate the coding mode. Another implementation determines a multiplier, to be used in a rate-distortion cost, based on pixel values from one or more of a video picture from a particular view or a rendered video picture for a different view. | 06-07-2012 |
20120200669 | FILTERING AND EDGE ENCODING - Several implementations relate, for example, to depth encoding and/or filtering for 3D video (3DV) coding formats. A sparse dyadic mode for partitioning macroblocks (MBs) along edges in a depth map is provided as well as techniques for trilateral (or bilateral) filtering of depth maps that may include adaptive selection between filters sensitive to changes in video intensity and/or changes in depth. One implementation partitions a depth picture, and then refines the partitions based on a corresponding image picture. Another implementation filters a portion of a depth picture based on values for a range of pixels in the portion. For a given pixel in the portion that is being filtered, the filter weights a value of a particular pixel in the range by a weight that is based on one or more of location distance, depth difference, and image difference. | 08-09-2012 |
20120206440 | Method for Generating Virtual Images of Scenes Using Trellis Structures - An image for a virtual view of a scene is generated based on a set of texture images and a corresponding set of depth images acquired of the scene. A set of candidate depth values associated with each pixel of a selected image is determined. For each candidate depth value, a cost that estimates a synthesis quality of the virtual image is determined. The candidate depth value with a least cost is selected to produce an optimal depth value for the pixel. Then, the virtual image is synthesized based on the optimal depth value of each pixel and the texture images. | 08-16-2012 |
20120249751 | IMAGE PAIR PROCESSING - At least one implementation determines whether two cameras are in parallel or are converging, based on an automated analysis of images from the cameras. One particular implementation determines the disparity of a foreground point and a background point. If the sign of the two disparities are the same, then the particular implementation decides that the cameras are in parallel. Otherwise, the particular implementation decides that the two cameras are converging. More generally, various implementations access a first image and a second image that form a stereo image pair. Multiple features are selected that exist in the first image and in the second image. An indicator of depth is determined for each of the multiple features. It is determined whether the first camera and the second camera were arranged in a parallel arrangement or a converging arrangement based on the values of the determined depth indicators. | 10-04-2012 |
20120249869 | STATMUX METHOD FOR BROADCASTING - A statistical multiplexing method is provided that comprises accessing a plurality of video sequences, wherein the video sequences are each assigned to a unique channel in a common broadcast system; collecting information from a plurality of the unique channels assigned to encode the corresponding video sequences; applying rho-domain analysis to the video sequences; and determining bitrate allocation for the channels responsive to the information collect and the rho-domain analysis. | 10-04-2012 |
20130039597 | Comfort Noise and Film Grain Processing for 3 Dimensional Video - Noise, either in the form of comfort noise or film grain, is added to a three dimensional image in accordance with image depth information to reduce human sensitivity to coding artifacts, thereby improving subjective image quality. | 02-14-2013 |
20140184744 | DEPTH CODING - Various implementations address depth coding and related disciplines. In one particular implementation, a segmentation is determined for a particular portion of a video image in a sequence of video images. The segmentation is determined based on reference depth indicators that are associated with at least a portion of one video image in the sequence of video images. Target depth indicators associated with the particular portion of the video image are processed. The processing is based on the determined segmentation in the particular portion of the video image. In another particular implementation, a segmentation is determined for at least a given portion of a video image based on depth indicators associated with the given portion. The segmentation is extended from the given portion into a target portion of the video image based on pixel values in the given portion and on pixel values in the target portion. | 07-03-2014 |