Patent application number | Description | Published |
20080238942 | Object-Based Image Inpainting - Systems and methods provide object-based image inpainting. The object-based inpainting can recover part of a missing object or can synthesize visual objects that are entirely missing. In one implementation, an exemplary system generates a target visual object to be inpainted into an image, by geometrically matching key features with a model visual object that can be selected from the same or another image. Borders of the target object are refined through pixel-based contour deformation and textures for the target object are obtained either by pixel interpolation or by patch-based modeling of pre-existing texture in the image. Graph-cut segmentation rectifies image overlap as the target object is embedded in the image. Poisson editing filters visual seams, and edge filling completes visual gaps around the border of the target object. | 10-02-2008 |
20080291065 | Wyner-Ziv Coding with Multiple Side Information - A Wyner-Ziv coding scheme jointly utilizes multiple side information at the decoder is described. A new probability model is introduced based on an extended turbo decoding trellis, so that the transitional probability is calculated from the probabilities conditioned on multiple side information. The proposed scheme handles multiple side information in a multi-dimensional manner to potentially achieve coding performance better than either selecting one side information or generating a single side information by averaging several side information. When the decoder receives multiple streams of parity bits, the conditional probabilities on the parity bits from the multiple streams all contribute in turbo decoding. The errors existing in some of the streams can be automatically eliminated during the decoding process. | 11-27-2008 |
20080292005 | Enhancement layer switching for scalable video coding - An exemplary system includes a data encoder generating a base layer bitstream encoded at a base bit-rate, and a plurality of enhancement layer bitstreams encoded at different enhancement layer bit-rates, and a bitstream selection module selecting one of the enhancement layer bitstreams every video frame based on available channel bandwidth. A method includes transmitting a first enhancement layer bitstream encoded at a first bit-rate, detecting a transition in network bandwidth through a switching bit-rate, and transmitting a second enhancement layer bitstream encoded at a second bit-rate based on the transition in network bandwidth. | 11-27-2008 |
20080310506 | Joint Spatio-Temporal Prediction for Video Coding - Systems and methods are described for joint spatio-temporal prediction (JSTP) during video coding. A given block of pixels is simultaneously predicted via inter frame prediction and intra frame prediction. The joint spatio-temporal prediction provides a new JSTP coding mode besides inter and intra modes in coding schemata such as the H.264/AVC video coding standard. In one implementation, subband decomposition obtains a high-pass component of inter frame prediction of a block and a low-pass component of intra frame prediction of the same block. The high-pass component of the inter prediction and the low-pass component of the intra prediction are combined to produce the joint spatio-temporal prediction. In one implementation, a difference between the original block and the JSTP prediction is coded in a bitstream to represent the block. | 12-18-2008 |
20090003430 | Video Coding Using Spatio-Temporal Texture Synthesis - Systems and methods for video coding using spatio-temporal texture synthesis are described. In one aspect, a video data coding pipeline portion of the codec removes texture blocks from the video data to generate coded video data. The removed texture blocks are selected based on an objective determination that each of the remove texture blocks can be synthesized from spatio-temporal neighboring samples during decoding operations. The objective determinations are made using local block-based motion information independent of global motion models. An indication of which texture blocks were removed is provided to a decoder in addition to the coded video data. Decoding logic of the codec decodes the video data using a standard decoding algorithm. The decoding logic also restores the removed texture blocks via spatio-temporal texture synthesis to generate synthesized video data. The decoded and synthesized video data is presented to a user. | 01-01-2009 |
20090003720 | EFFICIENT IMAGE REPRESENTATION BY EDGES AND LOW-RESOLUTION SIGNAL - An exemplary method for encoding an image includes receiving image data, detecting edges in the image data, selecting at least some of the detected edges, encoding the selected edges as selected edge information, down-sampling the image data, encoding the down-sampled image as down-sampled image information and multiplexing the selected edges information and the down-sampled image information. In such a method, the selected edges information and the down-sampled image information can be stored as an encoded image file. Other exemplary methods, devices, systems are also disclosed. | 01-01-2009 |
20090067491 | Learning-Based Image Compression - Learning-based image compression is described. In one implementation, an encoder possessing a first set of learned visual knowledge primitives excludes visual information from an image prior to compression. A decoder possessing an independently learned set of visual knowledge primitives synthesizes the excluded visual information into the image after decompression. The encoder and decoder are decoupled with respect to the information excluded at the encoder and the information synthesized at the decoder. This results in superior data compression since the information excluded at the encoder is dropped completely and not transferred to the decoder. Primitive visual elements synthesized at the decoder may be different than primitive visual elements dropped at the encoder, but the resulting reconstituted image is perceptually equivalent to the original image. | 03-12-2009 |
20090074307 | Textual Image Coding - Textual image coding involves coding textual portions of an image. In an example embodiment, a textual block of an image is decomposed into multiple base colors and an index map, with the index map having index values that each reference a base color so as to represent the textual block. A set of neighbor index values are ascertained for a particular index of the index map. A context that matches the neighbor index values is generated from among multiple contexts. The matching context includes a set of symbols. At least one symbol-to-value mapping is determined based on the matching context and a symbol to which the particular index corresponds. The particular index is remapped to a particular value in accordance with the symbol-to-value mapping and the symbol to which the particular index corresponds. | 03-19-2009 |
20090103606 | Progressive Distributed Video Coding - Progressive distributed video coding is described. In one implementation, video data maybe encoded by arranging the data into bit-planes. The arrangement of bit-planes is adapted by shifting the first non-zero bit-plane left by one place in the binary digits and moving the sign bit immediately in the place vacated by the shifted non-zero bit-plane. The adapted bit-planes are then encoded using an asymmetric Slepian-Wolf encoder. | 04-23-2009 |
20090238267 | Methods And Apparatuses For Use In Switching Between Streaming Video Bitstreams - Improved methods and apparatuses are provided for switching of streaming data bitstreams, such as, for example, used in video streaming and other related applications. Some desired functionalities provided herein include random access, fast forward and fast backward, error-resilience and bandwidth adaptation. The improved methods and apparatuses can be configured to increase coding efficiency of and/or reduce the amount of data needed to encode a switching bitstream. | 09-24-2009 |
20090238476 | Spectral information recovery for compressed image restoration - Aspects of the subject matter described herein relate to image restoration for compressed images. In aspects, image restoration is accomplished by recovering spectral information from data corresponding to a compressed image. The spectral information is recovered using an algorithm to search through a solution space of possible solutions while constraints are imposed on the solution space to trim undesirable solutions from the space. An algorithm described herein may be iteratively applied to improve the quality of the recovered image. | 09-24-2009 |
20090238484 | LIFTING-BASED DIRECTIONAL LAPPED TRANSFORMS - A “directional lapped transform coder” (“DLT Coder”) provides various techniques for designing directional lapped transforms. In general, lapped transforms are factorized into lifting steps. A “directional operator” is then introduced into each lifting step in order to construct the directional lapped transform by selecting data elements from a data signal along a path corresponding to the directional operator. The resulting directional lapped transform preserves the advantages of conventional lapped transforms while also providing more efficient representation directional signals. In various embodiments, the resulting directional lapped transform is used to enable an image/video coding scheme that provides significant improvement over conventional state-of-the-art image lapped transform based coding schemes for images with strong directional correlations. | 09-24-2009 |
20090245373 | VIDEO CODING - Improved video coding is described to encode video data within a sequence of video frames. To this end, at least a portion of a reference frame is encoded to include motion information associated with the portion of the reference frame. At least a portion of a predictable frame that includes video data predictively correlated to said portion of said reference frame is defined based on the motion information. At least said portion of the predictable frame is encoded without including corresponding motion information and including mode identifying data. The mode identifying data indicate that the encoded portion of the predictable frame can be directly derived using at least the motion information associated with the portion of the reference frame. | 10-01-2009 |
20090252146 | CONTINUOUS NETWORK CODING IN WIRELESS RELAY NETWORKS - Described is continuous network coding, in which a relay sends probability data comprising a continuous number for use as parity data. The node receives streams of bits sent from sources towards a destination, and computes the probability data based on current noise data and/or fading data. A selected set of the bits (all or some subset thereof) are combined, e.g., XOR-ed or concatenated, and send to the destination. Phase modulation is performed to convey probability information based on the probability data. The destination demodulates the signal to obtain the probability information, and combines the probability information with the data directly received from sources to perform joint decoding. The number of bits in the set of selected bits may be adaptively chosen based on current channel conditions, e.g., increased when the channel conditions from the sources directly to a destination are poor relative to the channel conditions via the relay. | 10-08-2009 |
20090252431 | Image Resizing for Web-based Image Search - Image resizing for web-based searching is described. In one implementation, a system resizes a user-selected thumbnail image into a larger version of the image that emulates the quality of a large, original image, but without downloading the original image. First, the system extracts resizing parameters when each thumbnail image is created. Then, the system creates a codebook of primitive visual elements extracted from a collection of training images. The primitive visual elements in the codebook provide universal visual parts for reconstructing images. The codebook and a resizing plug-in can be sent once to the user over a background channel. When the user selects a thumbnail image for enlargement, the system resizes the thumbnail image via interpolation and then refines the enlarged image with primitive visual elements from the codebook. The refinement creates an enlarged image that emulates the quality of the large, original image, without downloading the original image. | 10-08-2009 |
20090279462 | STATEFUL SCHEDULING WITH NETWORK CODING FOR ROADSIDE-TO-VEHICLE COMMUNICATION - Described is a technology by which a roadside-to-vehicle communication system may be implemented, including via a stateful scheduling with network coding scheme that enhances network capacity. Moving vehicles request and receive data from a roadside access points. Each of the access points operate a stateful scheduling algorithm that serves multiple vehicles by integrating network coding within a timeslot. In one aspect, the state of each vehicle's previously received and retained data is obtained, and used to enhance network capacity by combining as many packets as possible for multiple recipients in network coding. | 11-12-2009 |
20090285148 | NATURAL NETWORK CODING FOR MULTI-HOP WIRELESS NETWORK - Described herein is technology for, among other things, natural network coding in a wireless mesh network. The technology involves wireless mesh network systems, methods and devices based on the natural network coding. By encoding signals in their natural forms using their channel strengths, more efficient transmission of signals is possible in the wireless mesh network. | 11-19-2009 |
20090296670 | PULL-BASED DATA TRANSMISSION APPROACH - A network can include a number of nodes that link a source node to a sink node. When a first node in a network sends a packet to its downstream node, this information is also received at its upstream node. In response to learning that the first node has sent a packet, the upstream node sends another packet to the first node. In essence, a pull-based transmission approach is used to mitigate congestion and address the funneling effect in data transmission networks such as wireless video sensor networks. | 12-03-2009 |
20090304298 | HIGH DYNAMIC RANGE TEXTURE COMPRESSION - A method for compressing a high dynamic range (HDR) texture. A first block of texels of the HDR texture in a red-green-blue (RGB) space may be transformed to a second block of texels in a luminance-chrominance space. The first block may have red values, green values and blue values. The second block may have luminance values and chrominance values. The chrominance values may be based on a sum of the red values, a sum of the green values and a sum of the blue values. The luminance values and the chrominance values may be converted to an 8-bit integer format. The luminance values may be modified to restore a local linearity property to the second block. The second block may be compressed. | 12-10-2009 |
20090315905 | LAYERED TEXTURE COMPRESSION ARCHITECTURE - Various technologies for a layered texture compression architecture. In one implementation, the layered texture compression architecture may include a texture consumption pipeline. The texture compression pipeline may include a processor, memory devices, and textures compressed at varying ratios of compression. The textures within the pipeline may be compressed at ratios in accordance with characteristics of the devices in the pipeline that contains and processes the textures. | 12-24-2009 |
20090322777 | UNIFIED TEXTURE COMPRESSION FRAMEWORK - A method for compressing textures. A first block of texels is transformed from a red-green-blue (RGB) space to a second block of texels in a luminance-chrominance space. The first block has red values, green values and blue values. The second block has luminance values and chrominance values. The chrominance values may be based on a sum of the red values, a sum of the green values and a sum of the blue values. The chrominance values may be sampled for a first subset of texels in the second block. The luminance values and the sampled chrominance values may be converted to an 8-bit integer format. The luminance values of the first subset may be modified to restore a local linearity property to the first subset. The second block may be compressed into a third block. | 12-31-2009 |
20100020882 | Barbell Lifting for Wavelet Coding - A method for encoding motion-compensated video data includes generating, for a current frame, a high-pass wavelet coefficient based on a function of pixels in a temporally adjacent frame. The operations are repeated for multiple pixels in an array of pixels in the current frame to form an array of high-pass wavelet coefficients. A low-pass wavelet coefficient is generated based on a function of the high-pass wavelet coefficients. A system for coding video data includes a temporal wavelet decomposition module decomposing a pixel into a high-pass coefficient by performing a discrete wavelet transform on the pixel, a function of pixels in a previous frame, and/or a function of pixels in a subsequent frame. The system includes a motion estimation module generating motion vectors associated with the pixels in the previous frame and in the subsequent frame. | 01-28-2010 |
20100046849 | Image Restoration By Vector Quantization Utilizing Visual Patterns - The restoration of images by vector quantization utilizing visual patterns is disclosed. One disclosed embodiment comprises restoring detail in a transition region of an unrestored image, by first identifying the transition region and forming blurred visual pattern blocks. These blurred visual pattern blocks are compared to a pre-trained codebook, and a corresponding high-quality visual pattern blocks is obtained. The high-quality visual pattern block is then blended with the unrestored image to form a restored image. | 02-25-2010 |
20100074549 | IMAGE UPSAMPLING WITH TRAINING IMAGES - Images are upsampled using a knowledge base derived from a plurality of high-quality training images. The knowledge base is used to refine a high-frequency component including high-frequency aspects of a high-resolution, low-frequency image, interpolated from a low-resolution full-frequency image, into a high-frequency component. An enhancement step is performed without using a knowledge base to construct a high-compatibility component from the low-resolution, full-frequency image. The low-resolution, full-frequency image is combined with the coarse high-frequency component to yield an enhanced high-frequency component. A second knowledge base step is performed to construct an improved high-frequency component from the enhanced high-frequency component. The improved high-frequency component is blended with a high-resolution, low-frequency image to yield a high-resolution image. | 03-25-2010 |
20100080488 | FAST DIRECTIONAL IMAGE INTERPOLATOR WITH DIFFERENCE PROJECTION - Described is a technology in which a low resolution image is processed into a high-resolution image, including by a two interpolation passes. In the first pass, missing in-block pixels, which are the pixels within a block formed by four neighboring original pixels, are given values by gradient diffusion based upon interpolation of the surrounding original pixels. In the second interpolation pass, missing on-block pixels, which are the pixels on a block edge formed by two adjacent original pixels, are given values by gradient diffusion based upon interpolation of the values of those adjacent original pixels and the previously interpolated values of their adjacent in-block pixels. Also described is a difference projection process that varies the values of the interpolated pixels according to a computed difference projection. | 04-01-2010 |
20100135390 | VIDEO CODING - Improved video coding is described to encode video data within a sequence of video frames. To this end, at least a portion of a reference frame is encoded to include motion information associated with the portion of the reference frame. At least a portion of a predictable frame that includes video data predictively correlated to said portion of said reference frame is defined based on the motion information. At least said portion of the predictable frame is encoded without including corresponding motion information and including mode identifying data. The mode identifying data indicate that the encoded portion of the predictable frame can be directly derived using at least the motion information associated with the portion of the reference frame. | 06-03-2010 |
20100158400 | Accelerated Screen Codec - An accelerated screen codec technique is described that provides a general screen compression framework, which, in one embodiment, is Graphics Processor Unit (GPU) friendly. In one embodiment, in order to compress screen data, blocks in a compound screen image containing both images and text are segmented into text blocks and pictorial blocks using a simple gradient-based procedure. The text and pictorial blocks are then compressed respectively via different compression techniques. Additionally, a GPU acceleration architecture of one embodiment of the accelerated screen codec technique provides a screen codec that maximally exploits a GPU's high parallelism and reduces the download bandwidth from GPU to Computer Processing Unit (CPU). | 06-24-2010 |
20110170801 | RESIZING OF DIGITAL IMAGES - Digital images are resized according to a prescribed image scaling factor. An original image is re-sampled according to the scaling factor, resulting in an initial resized image. A probability of text (POT) map is generated for the initial resized image, where the POT map specifies a smoothed POT value for each pixel in the initial resized image. A weighting factor (WF) map is generated which maps each different smoothed POT value to a particular WF value. The WF map is used to calculate an adjusted luminance value for each pixel in the initial resized image, resulting in a final resized image. | 07-14-2011 |
20110191496 | COMPRESSIVE DATA GATHERING FOR LARGE-SCALE WIRELESS SENSOR NETWORKS - Techniques for data gathering in large-scale wireless sensor networks are described. A data collection device receives aggregate data from at least one sensor node of a group of N sensor nodes. The aggregate data includes M weighted sums. Each of the M weighted sums includes a respective sum of N products each of which being a product of a respective coefficient and a sensor reading from a respective one of the N sensor nodes. M and N are positive integers and M is less than N. Computation is performed on the aggregate data to recover sensor readings from the N sensor nodes. | 08-04-2011 |
20110262049 | VISION-BASED COMPRESSION - Systems and methods provide vision-based image compression. In one implementation, inpainting is the vision-based technique selected to augment a conventional signal-processing-based technique. For some regions of a source image, an exemplary system efficiently extracts and organizes structural edge information instead of compressing the regions. In one implementation, the system applies binary curve fitting to capture the edge information. A structure-aware inpainter in the decoder can then restore the regions via the edge information, which occupies very little data space or minimal bandwidth in a bitstream that is transmitted from encoder to decoder. Key visual components of the image can still be conventionally compressed. Extracting edge information for some regions instead of compressing them considerably increases overall image compression. | 10-27-2011 |
20120027313 | LINE AND PIXEL BASED METHODS FOR INTRA FRAME CODING - Images can be coded based on filters defined by filter coefficients or weights assigned to previously decoded pixel values of nearby lines. The filter coefficients can be selected based on evaluation of a set of predetermined filters, or filters can be adaptively generated and then evaluated. The filters generally are based on pixel values in previous rows or columns, but not values in the current row or column. In some examples, filters are adaptively generated pixel by pixel, or portions of previous rows, columns, or previously decoded portions of the current row or columns are used to provide line templates. A filter is generally selected to reduce coding error based on a least square difference between the current value and a prediction. | 02-02-2012 |
20120045118 | IMAGE RESIZING FOR WEB-BASED IMAGE SEARCH - Image resizing for web-based searching is described. In one implementation, a system resizes a user-selected thumbnail image into a larger version of the image that emulates the quality of a large, original image, but without downloading the original image. First, the system extracts resizing parameters when each thumbnail image is created. Then, the system creates a codebook of primitive visual elements extracted from a collection of training images. The primitive visual elements in the codebook provide universal visual parts for reconstructing images. The codebook and a resizing plug-in can be sent once to the user over a background channel. When the user selects a thumbnail image for enlargement, the system resizes the thumbnail image via interpolation and then refines the enlarged image with primitive visual elements from the codebook. The refinement creates an enlarged image that emulates the quality of the large, original image, without downloading the original image. | 02-23-2012 |
20120076414 | External Image Based Summarization Techniques - Techniques involve visually summarizing documents (e.g., search results, a collection of documents, etc.) using images which are visually representative of the documents for which the images represent. The images representing the documents may be external images obtained from sources other than the documents. The external images may be obtained from the sources other than the documents by performing a separate image based search using key phrases from the documents rather than extracting the images directly from within the documents themselves. Alternatively, an algorithm may be used to determine an image type, which may be chosen from a selection of external images, thumbnail images, or internal imaged taken directly from the collection of documents, that is suited to represent each document in the collection of documents. A snippet of the documents may be displayed along with the images which visually represent each of the documents. | 03-29-2012 |
20120082219 | Content adaptive deblocking during video encoding and decoding - Disclosed herein are exemplary embodiments of methods, apparatus, and systems for performing content-adaptive deblocking to improve the visual quality of video images compressed using block-based motion-predictive video coding. For instance, in certain embodiments of the disclosed technology, edge information is obtained using global orientation energy edge detection (“OEED”) techniques on an initially deblocked image. OEED detection can provide a robust partition of local directional features (“LDFs”). For a local directional feature detected in the partition, a directional deblocking filter having an orientation corresponding to the orientation of the LDF can be used. The selected filter can have a filter orientation and activation thresholds that better preserve image details while reducing blocking artifacts. In certain embodiments, for a consecutive non-LDF region, extra smoothing can be imposed to suppress the visually severe blocking artifacts. | 04-05-2012 |
20120121030 | Compressive Wireless Modulation - A projection code is applied to encode symbols as weighted arithmetic sums of approximately random subsets of binary source bits. Pairs of the symbols are combined to form constellation points, which are sequentially mapped through a constellation to modulate a data signal. | 05-17-2012 |
20120189056 | Video Coding Redundancy Reduction - Embodiments for reducing redundancy in video coding are disclosed. In accordance with at least one embodiment, video content is represented as a tree structure in which the nodes of the tree structure are associated with attributes of the video content. An encoder analyzes the nodes to determine whether there is redundancy associated with a last node of a divided node. In another embodiment, bidirectional prediction video coding techniques are applied to the video content. The encoder analyzes two lists of reference frames to determine whether a third mode of bidirectional prediction is redundant to a second mode of bidirectional prediction. The efficiency of video coding is improved by refraining from sending data symbols to the decoder in the event that the video coding techniques contain redundancy. | 07-26-2012 |
20120242674 | HIGH DYNAMIC RANGE TEXTURE COMPRESSION - A method for compressing a high dynamic range (HDR) texture. A first block of texels of the HDR texture in a red-green-blue (RGB) space may be transformed to a second block of texels in a luminance-chrominance space. The first block may have red values, green values and blue values. The second block may have luminance values and chrominance values. The chrominance values may be based on a sum of the red values, a sum of the green values and a sum of the blue values. The luminance values and the chrominance values may be converted to an 8-bit integer format. The luminance values may be modified to restore a local linearity property to the second block. The second block may be compressed. | 09-27-2012 |
20120287999 | SYNTAX ELEMENT PREDICTION IN ERROR CORRECTION - Architecture that improves error robustness in video coding and decoding. In particular, this can apply to motion vector prediction (MVP) such as a temporal MVP (TMVP). Flags can be used to indicate the use or non-use of a feature, such as to indicate whether the current slice uses or does not use TMVP, and to indicate in the slice header whether list prediction is allowed or not allowed. A flag can be signaled in sequence parameter set (SPS) or picture parameter set (PPS) as a way to enable an entire sequence to use or not use TMVP. TVMP can also be used to copy all the reference motion information to the current block. To address possible error problems, the full index of the TMVP can be recorded, and temporal information decoding refresh (TIDR) can be inserted into slices periodically. | 11-15-2012 |
20130041627 | Compressive Data Gathering for Large-Scale Wireless Sensor Networks - Techniques for data gathering in large-scale wireless sensor networks are described. A data collection device receives aggregate data from at least one sensor node of a group of N sensor nodes. The aggregate data includes M weighted sums. Each of the M weighted sums includes a respective sum of N products each of which being a product of a respective coefficient and a sensor reading from a respective one of the N sensor nodes. M and N are positive integers and M is less than N. Computation is performed on the aggregate data to recover sensor readings from the N sensor nodes. | 02-14-2013 |
20130051452 | VIDEO ENCODING ENHANCEMENTS - Techniques for implementing video encoding enhancements may increase video encoding efficiency. One of the techniques may involve the use of an exception value and a maximum index value in inter-prediction coding lookup table to enhance the combined coding of an inter-prediction direction and a reference frame index for a coding unit. Another of the techniques may include adaptively ordering the encoding of a split flag value and a skip flag value for a coding unit. An additional technique may include providing a uniform approach to adaptively combined code the coded block flag (CBF) values for transform or coding units of images that are used with the Residue Quad-tree (RQT) information to represent residue information. | 02-28-2013 |
20130129197 | IMAGE RESTORATION BY VECTOR QUANTIZATION UTILIZING VISUAL PATTERNS - The restoration of images by vector quantization utilizing visual patterns is disclosed. One disclosed embodiment comprises restoring detail in a transition region of an unrestored image, by first identifying the transition region and forming blurred visual pattern blocks. These blurred visual pattern blocks are compared to a pre-trained codebook, and a corresponding high-quality visual pattern blocks is obtained. The high-quality visual pattern block is then blended with the unrestored image to form a restored image. | 05-23-2013 |
20130148737 | SPATIOTEMPORAL PREDICTION FOR BIDIRECTIONALLY PREDICTIVE (B) PICTURES AND MOTION VECTOR PREDICTION FOR MULTI-PICTURE REFERENCE MOTION COMPENSATION - Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic. | 06-13-2013 |
20130208798 | TIMESTAMP-INDEPENDENT MOTION VECTOR PREDICTION FOR PREDICTIVE (P) AND BIDIRECTIONALLY PREDICTIVE (B) PICTURES - Methods and apparatuses are provided for achieving improved video coding efficiency through the use of Motion Vector Predictors (MVPs) for the encoding or decoding of motion parameters within the calculation of the motion information in B pictures and/or P pictures. Certain exemplary methods and apparatuses selectively apply temporal and/or spatial prediction. Rate Distortion Optimization (RDO) techniques are also applied in certain methods and apparatuses to further help improve coding efficiency. | 08-15-2013 |
20130223533 | VIDEO CODING - Improved video coding is described to encode video data within a sequence of video frames. To this end, at least a portion of a reference frame is encoded to include motion information associated with the portion of the reference frame. At least a portion of a predictable frame that includes video data predictively correlated to said portion of said reference frame is defined based on the motion information. At least said portion of the predictable frame is encoded without including corresponding motion information and including mode identifying data. The mode identifying data indicate that the encoded portion of the predictable frame can be directly derived using at least the motion information associated with the portion of the reference frame. | 08-29-2013 |
20140002445 | DEPTH SENSING WITH DEPTH-ADAPTIVE ILLUMINATION | 01-02-2014 |
20140037204 | Image Compression Based On Parameter-Assisted Inpainting - Systems and methods provide image compression based on parameter-assisted inpainting. In one implementation of an encoder, an image is partitioned into blocks and the blocks classified as smooth or unsmooth, based on the degree of visual edge content and chromatic variation in each block. Image content of the unsmooth blocks is compressed, while image content of the smooth blocks is summarized by parameters, but not compressed. The parameters, once obtained, may also be compressed. At a decoder, the compressed image content of the unsmooth blocks and the compressed parameters of the smooth blocks are each decompressed. Each smooth block is then reconstructed by inpainting, guided by the parameters in order to impart visual detail from the original image that cannot be implied from the image content of neighboring blocks that have been decoded. | 02-06-2014 |
20140086308 | Switching Between Streaming Video Bitstreams - Improved methods and apparatuses are provided for switching of streaming data bitstreams, such as, for example, used in video streaming and other related applications. Some desired functionalities provided herein include random access, fast forward and fast backward, error-resilience and bandwidth adaptation. The improved methods and apparatuses can be configured to increase coding efficiency of and/or reduce the amount of data needed to encode a switching bitstream. | 03-27-2014 |
20140105274 | Video Coding Redundancy Reduction - Embodiments for reducing redundancy in video coding are disclosed. In accordance with at least one embodiment, video content is represented as a tree structure in which the nodes of the tree structure are associated with attributes of the video content. An encoder analyzes the nodes to determine whether there is redundancy associated with a last node of a divided node. In another embodiment, bidirectional prediction video coding techniques are applied to the video content. The encoder analyzes two lists of reference frames to determine whether a third mode of bidirectional prediction is redundant to a second mode of bidirectional prediction. The efficiency of video coding is improved by refraining from sending data symbols to the decoder in the event that the video coding techniques contain redundancy. | 04-17-2014 |
20140254936 | LOCAL FEATURE BASED IMAGE COMPRESSION - The use of local feature descriptors of an image to generate compressed image data and reconstruct the image using image patches that are external to the image based on the compressed image data may increase image compression efficiency. A down-sampled version of the image is initially compressed to produce an encoded visual descriptor. The local feature descriptors of the image and the encoded visual descriptor are then obtained. A set of differential feature descriptors are subsequently determined based on the differences between the local feature descriptors of the input image and the encoded visual descriptor. At least some of the differential feature descriptors are compressed to produce encoded feature descriptors, which are then combined with the encoded visual feature descriptor produce image data. The image data may be used to select image patches from an image database to reconstruct the image. | 09-11-2014 |
20140313292 | DETERMINING DEPTH DATA FOR A CAPTURED IMAGE - A method, system, and one or more computer-readable storage media for depth acquisition from density modulated binary patterns are provided herein. The method includes capturing a number of images for a scene using an IR camera and a number of IR lasers including diffraction grates. Each image includes a density modulated binary pattern carrying phase information. The method also includes performing pixel based phase matching for the images to determine depth data for the scene based on the phase information carried by the density modulated binary patterns. | 10-23-2014 |
20140314316 | IMAGE COMPRESSION BASED ON PARAMETER-ASSISTED INPAINTING - Systems and methods provide image compression based on parameter-assisted inpainting. In one implementation of an encoder, an image is partitioned into blocks and the blocks classified as smooth or unsmooth, based on the degree of visual edge content and chromatic variation in each block. Image content of the unsmooth blocks is compressed, while image content of the smooth blocks is summarized by parameters, but not compressed. The parameters, once obtained, may also be compressed. At a decoder, the compressed image content of the unsmooth blocks and the compressed parameters of the smooth blocks are each decompressed. Each smooth block is then reconstructed by inpainting, guided by the parameters in order to impart visual detail from the original image that cannot be implied from the image content of neighboring blocks that have been decoded. | 10-23-2014 |
20140354681 | Absolute Phase Measurement with Secondary Pattern-Embedded Fringe - Multiple images of a scene are captured while a phase-shifted fringe embedded with a secondary pattern is projected onto the scene. A wrapped phase map is generated based on the captured images, and a continuous region mask is determined to segment the wrapped phase map into multiple continuous regions. A period disparity of the fringe is determined for each region based at least in part on the embedded secondary pattern, and the regional period disparities are applied to a spatially unwrapped phase map to produce an absolute phase map and finally a depth map. | 12-04-2014 |
20150016521 | VIDEO ENCODER FOR IMAGES - Some implementations can compress a digital image to create a compressed image corresponding to the digital image. The digital image can be divided into a plurality of coding units and a determination can be made whether to divide each coding unit of the plurality of coding units into smaller coding units/prediction units based on (i) a range of pixel values in each coding unit and/or (ii) a number of bits to encode the coding unit. Rate distortion optimization can be performed based on a size of each of the prediction units. A determination can be made whether to perform a transform for each of the transform units based on a size of each prediction unit and/or a spatial frequency measure associated with each transform unit. | 01-15-2015 |
20150016527 | SPATIOTEMPORAL PREDICTION FOR BIDIRECTIONALLY PREDICTIVE (B) PICTURES AND MOTION VECTOR PREDICTION FOR MULTI-PICTURE REFERENCE MOTION COMPENSATION - Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic. | 01-15-2015 |
20150023425 | CONTENT ADAPTIVE DEBLOCKING DURING VIDEO ENCODING AND DECODING - Disclosed herein are exemplary embodiments of methods, apparatus, and systems for performing content-adaptive deblocking to improve the visual quality of video images compressed using block-based motion-predictive video coding. For instance, in certain embodiments of the disclosed technology, edge information is obtained using global orientation energy edge detection (“OEED”) techniques on an initially deblocked image. OEED detection can provide a robust partition of local directional features (“LDFs”). For a local directional feature detected in the partition, a directional deblocking filter having an orientation corresponding to the orientation of the LDF can be used. The selected filter can have a filter orientation and activation thresholds that better preserve image details while reducing blocking artifacts. In certain embodiments, for a consecutive non-LDF region, extra smoothing can be imposed to suppress the visually severe blocking artifacts. | 01-22-2015 |
20150054923 | DEPTH SENSING WITH DEPTH-ADAPTIVE ILLUMINATION - An adaptive depth sensing system (ADSS) illuminates a scene with a pattern that is constructed based on an analysis of at least one prior-generated depth map. In one implementation, the pattern is a composite pattern that includes two or more component patterns associated with different depth regions in the depth map. The composite pattern may also include different illumination intensities associated with the different depth regions. By using this composite pattern, the ADSS can illuminate different objects in a scene with different component patterns and different illumination intensities, where those objects are located at different depths in the scene. This process, in turn, can reduce the occurrence of defocus blur, underexposure, and overexposure in the image information. | 02-26-2015 |