Patent application number | Description | Published |
20080216125 | Mobile Device Collaboration - Systems and methods are described for mobile device collaboration. An exemplary collaborative architecture enables aggregation of resources across two or more mobile devices, in such a manner that the aggregation of resources is practical even considering the miniaturized and limited battery power of most mobile devices. In a video implementation, the exemplary collaborative architecture senses when another mobile device is in close enough proximity to aggregate resources. The collaborative architecture applies an adaptive video decoder so that each mobile device can participate in playing back a larger and higher-resolution video across combined display screens than any single mobile device could playback alone. A cross-display motion prediction technique saves battery power by balancing the amount of collaborative communication between devices against the local processing that each device performs to display visual motion across the boundary separating displays. | 09-04-2008 |
20090003430 | Video Coding Using Spatio-Temporal Texture Synthesis - Systems and methods for video coding using spatio-temporal texture synthesis are described. In one aspect, a video data coding pipeline portion of the codec removes texture blocks from the video data to generate coded video data. The removed texture blocks are selected based on an objective determination that each of the remove texture blocks can be synthesized from spatio-temporal neighboring samples during decoding operations. The objective determinations are made using local block-based motion information independent of global motion models. An indication of which texture blocks were removed is provided to a decoder in addition to the coded video data. Decoding logic of the codec decodes the video data using a standard decoding algorithm. The decoding logic also restores the removed texture blocks via spatio-temporal texture synthesis to generate synthesized video data. The decoded and synthesized video data is presented to a user. | 01-01-2009 |
20090067491 | Learning-Based Image Compression - Learning-based image compression is described. In one implementation, an encoder possessing a first set of learned visual knowledge primitives excludes visual information from an image prior to compression. A decoder possessing an independently learned set of visual knowledge primitives synthesizes the excluded visual information into the image after decompression. The encoder and decoder are decoupled with respect to the information excluded at the encoder and the information synthesized at the decoder. This results in superior data compression since the information excluded at the encoder is dropped completely and not transferred to the decoder. Primitive visual elements synthesized at the decoder may be different than primitive visual elements dropped at the encoder, but the resulting reconstituted image is perceptually equivalent to the original image. | 03-12-2009 |
20090074307 | Textual Image Coding - Textual image coding involves coding textual portions of an image. In an example embodiment, a textual block of an image is decomposed into multiple base colors and an index map, with the index map having index values that each reference a base color so as to represent the textual block. A set of neighbor index values are ascertained for a particular index of the index map. A context that matches the neighbor index values is generated from among multiple contexts. The matching context includes a set of symbols. At least one symbol-to-value mapping is determined based on the matching context and a symbol to which the particular index corresponds. The particular index is remapped to a particular value in accordance with the symbol-to-value mapping and the symbol to which the particular index corresponds. | 03-19-2009 |
20090076882 | MULTI-MODAL RELEVANCY MATCHING - This document describes techniques capable of associating relevant entities, such as advertisements, with insertion points within a media file. These techniques calculate a global relevancy between entities and the media file. These techniques may also calculate a local relevancy between the entities and one or more insertion points within the media file. Both global and local relevancies may employ textual and non-textual information. With use of the calculated global and local relevancies, the techniques associate one or more entities with each of the one or more insertion points in the media file. These techniques thus enable, for each insertion point, associating a most relevant entity for a particular insertion point with the insertion point. Therefore, when a user consumes the media file the user may also consume a most relevant entity at and for each insertion point in the media file. | 03-19-2009 |
20090079871 | ADVERTISEMENT INSERTION POINTS DETECTION FOR ONLINE VIDEO ADVERTISING - Systems and methods for determining insertion points in a first video stream are described. The insertions points being configured for inserting at least one second video into the first video. In accordance with one embodiment, a method for determining the insertion points includes parsing the first video into a plurality of shots. The plurality of shots includes one or more shot boundaries. The method then determines one or more insertion points by balancing a discontinuity metric and an attractiveness metric of each shot boundary. | 03-26-2009 |
20090083010 | Correlative Multi-Label Image Annotation - Correlative multi-label image annotation may entail annotating an image by indicating respective labels for respective concepts. In an example embodiment, a classifier is to annotate an image by implementing a labeling function that maps an input feature space and a label space to a combination feature vector. The combination feature vector models both features of individual ones of the concepts and correlations among the concepts. | 03-26-2009 |
20090083781 | Intelligent Video Player - Systems and methods for managing digital video data are described. The digital video data maybe managed by employing a computing device to extract metadata from the video file and calculate a unique video signature associated with the video file. The computing device then uploads the metadata and unique video signature to a server which stores the metadata in a lookup table according to the unique video signature. | 03-26-2009 |
20090125461 | Multi-Label Active Learning - Multi-label active learning may entail training a classifier with a set of training samples having multiple labels per sample. In an example embodiment, a method includes accepting a set of training samples, with the set of training samples having multiple respective samples that are each respectively associated with multiple labels. The set of training samples is analyzed to select a sample-label pair responsive to at least one error parameter. The selected sample-label pair is then submitted to an oracle for labeling. | 05-14-2009 |
20090171787 | Impressionative Multimedia Advertising - A method for making online adverisement makes an impressionative presentation of an advertisement to a viewer. The impressionative presentation is an impressionized version of an original online source medium such as a photo. The method associates advertisements with the source medium based, at least in part, on calculated ad relevance, and determines one or more viewer iteractive points on the original source medium. The method then presents to the viewer an ad-augmented medium including an impressionized version of the source medium, which has the ability to change the form of impression to a viewer in response to an interactive act conducted by the viewer. The ad-augmented medium may include the associated advertisement content or direct the viewer's attention thereto. | 07-02-2009 |
20090274434 | VIDEO CONCEPT DETECTION USING MULTI-LAYER MULTI-INSTANCE LEARNING - Visual concepts contained within a video clip are classified based upon a set of target concepts. The clip is segmented into shots and a multi-layer multi-instance (MLMI) structured metadata representation of each shot is constructed. A set of pre-generated trained models of the target concepts is validated using a set of training shots. An MLMI kernel is recursively generated which models the MLMI structured metadata representation of each shot by comparing prescribed pairs of shots. The MLMI kernel is subsequently utilized to generate a learned objective decision function which learns a classifier for determining if a particular shot (that is not in the set of training shots) contains instances of the target concepts. A regularization framework can also be utilized in conjunction with the MLMI kernel to generate modified learned objective decision functions. The regularization framework introduces explicit constraints which serve to maximize the precision of the classifier. | 11-05-2009 |
20090313294 | AUTOMATIC IMAGE ANNOTATION USING SEMANTIC DISTANCE LEARNING - Images are automatically annotated using semantic distance learning. Training images are manually annotated and partitioned into semantic clusters. Semantic distance functions (SDFs) are learned for the clusters. The SDF for each cluster is used to compute semantic distance scores between a new image and each image in the cluster. The scores for each cluster are used to generate a ranking list which ranks each image in the cluster according to its semantic distance from the new image. An association probability is estimated for each cluster which specifies the probability of the new image being semantically associated with the cluster. Cluster-specific probabilistic annotations for the new image are generated from the manual annotations for the images in each cluster. The association probabilities and cluster-specific probabilistic annotations for all the clusters are used to generate final annotations for the new image. | 12-17-2009 |
20100106486 | IMAGE-BASED SEMANTIC DISTANCE - Image-based semantic distance technique embodiments are presented that involve establishing a measure of an image-based semantic distance between semantic concepts. Generally, this entails respectively computing a semantic concept representation for each concept based on a collection of images associated with the concept. A degree of difference is then computed between two semantic concept representations to produce the aforementioned semantic distance measure for the pair of corresponding concepts. | 04-29-2010 |
20100135390 | VIDEO CODING - Improved video coding is described to encode video data within a sequence of video frames. To this end, at least a portion of a reference frame is encoded to include motion information associated with the portion of the reference frame. At least a portion of a predictable frame that includes video data predictively correlated to said portion of said reference frame is defined based on the motion information. At least said portion of the predictable frame is encoded without including corresponding motion information and including mode identifying data. The mode identifying data indicate that the encoded portion of the predictable frame can be directly derived using at least the motion information associated with the portion of the reference frame. | 06-03-2010 |
20100149419 | MULTI-VIDEO SYNTHESIS - Embodiments that provide multi-video synthesis are disclosed. In accordance with one embodiment, multi-video synthesis includes breaking a main video into a plurality of main frames and break a supplementary video into a plurality of supplementary frames. The multi-video synthesis also includes assigning one or more supplementary frames into each of a plurality of states of a Hidden Markov Model (HMM), where each of the plurality of states corresponding to one or more main frames. The multi-video synthesis further includes determining optimal frames in the plurality of main frames for insertion of the plurality of supplementary frames based on the plurality of states and visual properties. The optimal frames include optimal insertion positions. The multi-video synthesis additionally includes inserting the plurality of supplementary frames into the optimal insertion positions to form a synthesized video. | 06-17-2010 |
20100158412 | INTERACTIVELY RANKING IMAGE SEARCH RESULTS USING COLOR LAYOUT RELEVANCE - This disclosure describes various exemplary user interfaces, methods, and computer program products for the interactively ranking image search results refinement method using a color layout. The method includes receiving a text query for an image search, presenting image search results in a structured presentation based on the text query and information from an interest color layout. The process creates image search results that may be selected by the user based on color selection palettes or color layout specification schemes. Then the process ranks the image search results by sorting the results according to similarity scores between color layouts from the image search results and the interest color layout from a user based on the color selection palettes and the color layout specification schemes. | 06-24-2010 |
20110075992 | INTELLIGENT OVERLAY FOR VIDEO ADVERTISING - Video advertising overlay technique embodiments are presented that generally detect a set of spatio-temporal nonintrusive positions within a series of consecutive video frames in shots of a digital video and then overlay contextually relevant ads on these positions. In one general embodiment, this is accomplished by decomposing the video into a series of shots, and then identifying a video advertisement for each of a selected set of the shots. The identified video advertisement is one that is determined to be the most relevant to the content of the shot. An overlay area is also identified in each of the shots, where the selected overlay area is the least intrusive among a plurality of prescribed areas to a viewer of the video. The video advertisements identified for the shots are then respectively scheduled to be overlaid in the identified overlay area of a shot, whenever the shot is played. | 03-31-2011 |
20110199389 | INTERACTIVE VIRTUAL DISPLAY SYSTEM FOR UBIQUITOUS DEVICES - An “Interactive Virtual Display,” as described herein, provides various systems and techniques that facilitate ubiquitous user interaction with both local and remote heterogeneous computing devices. More specifically, the Interactive Virtual Display uses various combinations of small-size programmable hardware and portable or wearable sensors to enable any display surface (e.g., computer display devices, televisions, projected images/video from projection devices, etc.) to act as a thin client for users to interact with a plurality heterogeneous computing devices regardless of where those devices are located relative to the user. The Interactive Virtual Display provides a flexible system architecture that enables communication and collaboration between a plurality of both local and remote heterogeneous computing devices. This communication and collaboration enables a variety of techniques, such as adaptive screen compression, user interface virtualization, real-time gesture detection to improve system performance and overall user experience, etc. | 08-18-2011 |
20110289015 | MOBILE DEVICE RECOMMENDATIONS - Users may browse web pages, interact with a plethora of applications, search for new content, and perform a wide variety of other tasks using a mobile device. Unfortunately, useful content may be difficult for a user to locate because of the large amount of content available (e.g. hundreds of thousands of applications within an application store). Accordingly, one or more systems and/or techniques for determining recommendations are disclosed herein. In particular, user input (e.g., text, numbers, etc.) and/or a user profile (e.g., contextual information relating to a user) may be used to determine a user intent. Recommendations may be determined based upon the user intent. For example, a user may input “I am hungry” using a mobile phone having a GPS location of Downtown and a noon timestamp. Using this information, an application allowing the user to make lunch reservations at local restaurants may be provided as a recommendation. | 11-24-2011 |
20120260157 | Cooperative Rendering Cache for Mobile Browser - A cooperative rendering cache browser (CRC-Browser) for a mobile device may cooperatively manage cached content with a proxy server to reduce redundant transmissions of processed Web data. Additionally, the CRC-Browser may provide stateful, thin-client Web browsing to maintain synchronized information about rendering objects of a Web page. Further, the CRC-Browser may cache only stable and/or visible document object model (DOM) elements and communicate a list of IDs of the cached elements to the proxy. Such stateful thin-client browsers may be useful for reducing battery drain, poor responsiveness, and high wireless network costs of mobile Web browsing. | 10-11-2012 |
20130013596 | DOCUMENT-RELATED REPRESENTATIVE INFORMATION - Some implementations include techniques and arrangements to provide document-related representative information with search results. For example, a representative query and/or representative results may be provided for one or more individual documents identified in a set of search results to supplement the search results returned in response to a received search query. The representative queries may be determined by correlating a plurality of previously submitted queries in search log data with a plurality of documents returned in response to the queries. In some implementations, click-through frequency for a particular document with respect to the plurality of queries may be taken into consideration when determining the representative queries for the particular document. In some implementations, the representative queries serve to categorize the search results based on subject matter, and a link may be provided to representative results corresponding to the representative query for accessing documents directed to similar subject matter. | 01-10-2013 |
20130014860 | Method for manufacturing melt-spinning alloys and apparatus thereofAANM Li; HongweiAACI BeijingAACO CNAAGP Li; Hongwei Beijing CNAANM Yu; DunboAACI BeijingAACO CNAAGP Yu; Dunbo Beijing CNAANM Luo; YangAACI BeijingAACO CNAAGP Luo; Yang Beijing CNAANM Li; KuosheAACI BeijingAACO CNAAGP Li; Kuoshe Beijing CNAANM Li; ShipengAACI BeijingAACO CNAAGP Li; Shipeng Beijing CNAANM Wang; MinAACI BeijingAACO CNAAGP Wang; Min Beijing CNAANM Yuan; YongqiangAACI BeijingAACO CNAAGP Yuan; Yongqiang Beijing CN - The application provides a method for manufacturing melt-spinning alloys and an apparatus thereof, which belongs to the technical field of metal materials and preparation thereof. The main feature of method including steps of melting alloy and jetting the molten alloy for rapid-quenching is that alloy melting and rapid-quenching are respectively implemented in independent environments, and the pressure of the two environments can be adjusted separately. The method can realize uniformity control of rapid-quenching velocity by controlling the melting and quenching pressure respectively, which has the advantages of increased rapid-quenching cooling rate, improved melt-spinning alloys thickness uniformity, reduced probability of nozzle clogging. | 01-17-2013 |
20130020527 | RARE-EARTH PERMANENT MAGNETIC POWDER, BONDED MAGNET, AND DEVICE COMPRISING THE SAME - A rare-earth permanent magnetic powder, a bonded magnet, and a device comprising the bonded magnet are provided. The rare-earth permanent magnetic powder is mainly composed of 7-12 at % of Sm, 0.1-1.5 at % of M, 10-15 at % of N, 0.1-1.5 at % of Si, and Fe as the balance, wherein M is at least one element selected from the group of Be, Cr, Al, Ti, Ga, Nb, Zr, Ta, Mo, and V, and the main phase of the rare-earth permanent magnetic powder is of TbCu | 01-24-2013 |
20130101014 | Layered Screen Video Encoding - A computing device is described herein that is configured to encode natural video content in accordance with a first encoding scheme and screen content in accordance with a second encoding scheme. The computing device is configured to distinguish between the natural video content of a video frame and the screen content of the video frame based at least in part on temporal correlations between the video frame and one or more neighboring video frames and on content analysis of the video frame. | 04-25-2013 |
20130121573 | HYBRID CODEC FOR COMPOUND IMAGE COMPRESSION - One or more portions of a compound image may be classified as picture portions and at least one remaining portion of the compound image may be classified as a non-picture portion. A first layer of a layered image may be generated based on the picture portions of the compound image. The first layer may be compliant with a first image format. A second layer of the layered image may be generated based on the non-picture portion. The second layer may be compliant with a second image format that is different from the first image format. The first layer and the second layer may be sent to a web browser. The first format and the second format may be supported by the web browser. | 05-16-2013 |
20130148737 | SPATIOTEMPORAL PREDICTION FOR BIDIRECTIONALLY PREDICTIVE (B) PICTURES AND MOTION VECTOR PREDICTION FOR MULTI-PICTURE REFERENCE MOTION COMPENSATION - Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic. | 06-13-2013 |
20130208798 | TIMESTAMP-INDEPENDENT MOTION VECTOR PREDICTION FOR PREDICTIVE (P) AND BIDIRECTIONALLY PREDICTIVE (B) PICTURES - Methods and apparatuses are provided for achieving improved video coding efficiency through the use of Motion Vector Predictors (MVPs) for the encoding or decoding of motion parameters within the calculation of the motion information in B pictures and/or P pictures. Certain exemplary methods and apparatuses selectively apply temporal and/or spatial prediction. Rate Distortion Optimization (RDO) techniques are also applied in certain methods and apparatuses to further help improve coding efficiency. | 08-15-2013 |
20130223533 | VIDEO CODING - Improved video coding is described to encode video data within a sequence of video frames. To this end, at least a portion of a reference frame is encoded to include motion information associated with the portion of the reference frame. At least a portion of a predictable frame that includes video data predictively correlated to said portion of said reference frame is defined based on the motion information. At least said portion of the predictable frame is encoded without including corresponding motion information and including mode identifying data. The mode identifying data indicate that the encoded portion of the predictable frame can be directly derived using at least the motion information associated with the portion of the reference frame. | 08-29-2013 |
20130223740 | Salient Object Segmentation - Techniques for identifying a salient object with respect to its context are described. A process receives an input image that includes a salient object. The process segments the input image into multiple regions and calculates a saliency value for each of the segmented regions based on scale image levels. The process constructs saliency maps based at least in part on the calculated saliency value, and combines the saliency maps to construct a total saliency map. Next, the process connects a set of line segments computed from the input image and utilizes the total saliency map to compute a closed boundary, which forms a shape prior from the closed boundary, and extracts the salient object from the total saliency map and the shape prior. | 08-29-2013 |
20130230255 | Image Searching By Approximate k-NN Graph - This disclosure describes techniques for searching for similar images to an image query by using an approximate k-Nearest Neighbor (k-NN) graph. The approximate k-NN graph is constructed from data points partitioned into subsets to further identify nearest-neighboring data points for each data point. The data points may connect with the nearest-neighboring data points in a subset to form an approximate neighborhood subgraph. These subgraphs from all the subsets are combined together to form a base approximate k-NN graph. Then by performing more random hierarchical partition, more base approximate k-NN graphs are formed, and further combined together to create an approximate k-NN graph. The approximate k-NN graph expands into other neighborhoods and identifies the best k-NN data points. The approximate k-NN graph retrieves the best NN data points, based at least in part on the retrieved best k-NN data points representing images being similar in appearance to the image query. | 09-05-2013 |
20140053090 | INTERACTIVE VIRTUAL DISPLAY SYSTEM - An “Interactive Virtual Display,” as described herein, provides various systems and techniques that facilitate ubiquitous user interaction with both local and remote heterogeneous computing devices. More specifically, the Interactive Virtual Display uses various combinations of small-size programmable hardware and portable or wearable sensors to enable any display surface (e.g., computer display devices, televisions, projected images/video from projection devices, etc.) to act as a thin client for users to interact with a plurality heterogeneous computing devices regardless of where those devices are located relative to the user. The Interactive Virtual Display provides a flexible system architecture that enables communication and collaboration between a plurality of both local and remote heterogeneous computing devices. This communication and collaboration enables a variety of techniques, such as adaptive screen compression, user interface virtualization, real-time gesture detection to improve system performance and overall user experience, etc. | 02-20-2014 |
20140055560 | Depth Data Processing and Compression - Techniques for setting depth values for invalid measurement regions of depth images are described herein. A computing device may set the depth values based on evaluations of depth values of neighboring pixels and of corresponding pixels from time-adjacent depth images. Alternately or additionally, the computing device may utilize a texture image corresponding to the depth image to identify objects and may set depth values for pixels based on depth values of other pixels belonging to the same object. After setting the depth values, the computing device may normalize the depth values of the pixels. Further, the computing device may generate reduced representations of the depth images based on a depth reference model or a depth error model and may provide the reduced representations to an encoder. | 02-27-2014 |
20140250109 | RERANKING USING CONFIDENT IMAGE SAMPLES - The techniques described herein determine an initial set of ranked images associated with an image-based search query. Based on visual content similarities between images in the initial set of ranked images, the techniques select confident image samples from the initial set of ranked images. The techniques then use the confident image samples to rerank the initial set of ranked images. Accordingly, a search engine uses the confident image samples to promote images that are likely to be relevant to the search query, while demoting images that are not likely to be relevant to the search query. Therefore, the search engine can provide improved relevance-based search results to an image-based search query. | 09-04-2014 |
20140250110 | IMAGE ATTRACTIVENESS BASED INDEXING AND SEARCHING - Attractiveness of an image may be estimated by integrating extracted visual features with contextual cues pertaining to the image. Image attractiveness may be defined by the visual features (e.g., perceptual quality, aesthetic sensitivity, and/or affective tone) of elements contained within the image. Images may be indexed based on the estimated attractiveness, search results may be presented based on image attractiveness, and/or a user may elect, after receiving image search results, to re-rank the image search results by attractiveness. | 09-04-2014 |
20140250120 | Interactive Multi-Modal Image Search - A facility for visual search on a mobile device takes advantage of multi-modal and multi-touch input on the mobile device. By extracting lexical entities from a spoken search query and matching the lexical entities to image tags, the facility provides candidate images for each entity. Selected ones of the candidate images are used to construct a composite visual query image on a query canvas. The relative size and position of the selected candidate images in the composite visual query image, which need not be an existing image, contribute to a definition of a context of the composite visual query image being submitted for context-aware visual search. | 09-04-2014 |
20150016527 | SPATIOTEMPORAL PREDICTION FOR BIDIRECTIONALLY PREDICTIVE (B) PICTURES AND MOTION VECTOR PREDICTION FOR MULTI-PICTURE REFERENCE MOTION COMPENSATION - Several improvements for use with Bidirectionally Predictive (B) pictures within a video sequence are provided. In certain improvements Direct Mode encoding and/or Motion Vector Prediction are enhanced using spatial prediction techniques. In other improvements Motion Vector prediction includes temporal distance and subblock information, for example, for more accurate prediction. Such improvements and other presented herein significantly improve the performance of any applicable video coding system/logic. | 01-15-2015 |