Patent application number | Description | Published |
20110064375 | IMAGE PROCESSING METHOD, APPARATUS AND SYSTEM - An image processing method comprises receiving a sequence of image frames from a video camera. The method includes generating a computer generated object, and combining the computer generated object with the sequence of image frames so as to generate a sequence of augmented reality images. The method further includes dividing each received image frame into a respective array of image motion cells, detecting inter-image motion in successive image frames for each motion cell, and generating a motion object comprising one or more image motion cells. The image motion cells in the motion object correspond to a set of image motion cells detected as comprising inter-image motion over a threshold number of image frames. The method includes detecting a relative distance between the computer generated object and the motion object within the augmented reality images, and generating a point of interest within a current image frame in dependence upon the detected distance between the computer generated object and the motion object so that the computer generated object can be caused to appear to interact with an image region corresponding to an image motion cell at the point of interest. | 03-17-2011 |
20110141301 | IMAGE PROCESSING METHOD, APPARATUS AND SYSTEM - An image processing method comprises receiving, from a camera, a camera captured image comprising a user drawn image. At least part of the user drawn image is within a field of view of the camera. The method comprises applying an image processing operation to the camera captured image so as to generate an image feature within the camera captured image which corresponds to the user drawn image, and thinning the image feature to a line width which is smaller than a line width associated with the image feature so as to generate a modified width image feature. The method further comprises generating vector data from pixels associated with the modified width image feature so that the vector data substantially represents the modified width image feature. | 06-16-2011 |
20120262485 | SYSTEM AND METHOD OF INPUT PROCESSING FOR AUGMENTED REALITY - A method of input processing for augmented reality comprises the steps of capturing a video image, generating an augmented image layer for superposition over the captured video image, and for a region of the augmented image layer, detecting for each pixel in the region a property of a corresponding pixel in the captured video image, and mapping with a first mapping the property detected for each pixel of the region back to a reference two-dimensional array of pixels; and generating an input based upon the property values as mapped to the reference two-dimensional array of pixels. | 10-18-2012 |
20120262486 | SYSTEM AND METHOD OF USER INTERACTION FOR AUGMENTED REALITY - A method of user interaction in augmented reality comprises the steps of capturing a video image of a scene, and for each pixel in at least a sub-region of the captured video, classifying the pixel as either a skin or non-skin pixel responsive to whether the colour of the pixel exceeds a predetermined threshold purity of red; and generating a mask based upon the classification of the pixels of the captured video, generating an augmentation image layer to superpose on the captured video image, and limiting a mode of combination of the captured video and the augmentation image layer, responsive to the mask. | 10-18-2012 |
20130040737 | INPUT DEVICE, SYSTEM AND METHOD - An input device for an entertainment device comprises a first main surface, one or more side edges, one or more motion sensors operable to generate motion data, and a wireless communicator for transmitting the motion data to the entertainment device. An elongate illuminated region extends along some or all of one or more of the side edges. In addition, the input device may comprise a first stereoscopic camera and a second stereoscopic camera having respective viewpoints located at opposing corners of a first area of the main surface. The field of view of each stereoscopic camera may encompassing a first volume of space immediately above the first area. The first and second stereoscopic cameras are operable to generate respective stereoscopic images, from which a first depth map and complementary second depth map may be generated for opposite sides of an object placed within the first volume of space. | 02-14-2013 |