# Kun Zhou, Beijing CN

## Kun Zhou, Beijing CN

Patent application number | Description | Published |
---|---|---|

20080303840 | Mesh Quilting for Geometric Texture Synthesis - Mesh quilting for geometric texture synthesis involves synthesizing a geometric texture by quilting a mesh texture swatch. In an example embodiment, geometry is matched between a mesh texture swatch and a portion of a synthesized geometric texture. Correspondences are ascertained between elements of the mesh texture swatch and the portion of the synthesized geometric texture. The ascertained corresponding elements of the mesh texture swatch and the portion of the synthesized geometric texture are aligned via local deformation to create a new patch. The new patch is merged into an output texture space to grow the synthesized geometric texture. | 12-11-2008 |

20080309664 | Mesh Puppetry - This disclosure describes a variational framework for detail-preserving skinned mesh manipulation or deformation. The skinned mesh deformation occurs by optimizing skeleton position and vertex weights of a skeletal skinned mesh in an integrated manner. The process allows creating new poses and animations by specifying a few desired constraints for the skeletal skinned mesh in an interactive deformation platform. This process adjusts the skeletal position and solves for a deformed skinned mesh simultaneously with an algorithm in conjunction with the constraints. The algorithm includes a cascading optimization procedure. The mesh puppetry displays skinned mesh manipulation in real-time. | 12-18-2008 |

20080309667 | Interactive Relighting with Dynamic Reflectance - Interactive relighting with dynamic reflectance involves relighting a graphical scene with dynamic changes to the reflectance(s) in the graphical scene. A graphical scene may include source radiance, regions having reflectances, a surface spot, incident radiation from the source radiance at the surface sport, an incident direction, a viewing direction, exit radiance, and so forth. In an example embodiment, a graphical scene is relighted based on at least one adjusted reflectance of the graphical scene using an incident radiance at a surface spot that is separated into respective incident radiance components corresponding to different respective numbers of interreflections in the graphical scene. In another example embodiment, a graphical scene is relighted based on at least one adjusted reflectance of the graphical scene using a tensor representation for a reflectance of a surface spot with the tensor representation being segmented into three adjustable factors for lighting, viewing, and reflectance. | 12-18-2008 |

20080316202 | DIRECT MANIPULATION OF SUBDIVISION SURFACES USING A GRAPHICS PROCESSING UNIT - A graphics system allows for manipulation of a detail mesh for a subdivision surface. To deform the subdivision surface, the graphics system generates a corresponding deformed control mesh by attempting to satisfy both position constraints of the manipulation and Laplacian constraints for the detail mesh. After the deformed control mesh is generated, the deformed detail mesh can be generated by applying a subdivision function to the deformed control mesh to generate a deformed smooth mesh and then applying detail information to the deformed smooth mesh. | 12-25-2008 |

20090002376 | Gradient Domain Editing of Animated Meshes - Gradient domain editing of animated meshes is described. Exemplary systems edit deforming mesh sequences by applying Laplacian mesh editing techniques in the spacetime domain. A user selects relevant frames or handles to edit and the edits are propagated to the entire sequence. For example, if the mesh depicts an animated figure, then user-modifications to position of limbs, head, torso, etc., in one frame are propagated to the entire sequence. In advanced editing modes, a user can reposition footprints over new terrain and the system automatically conforms the walking figure to the new footprints. A user-sketched curve can automatically provide a new motion path. Movements of one animated figure can be transferred to a different figure. Caricature and cartoon special effects are available. The user can also select spacetime morphing to smoothly change the shape and motion of one animated figure into another over a short interval. | 01-01-2009 |

20090006044 | Real-Time Rendering of Light-Scattering Media - A real-time algorithm for rendering of an inhomogeneous scattering media such as smoke under dynamic low-frequency environment lighting is described. An input media animation is represented as a sequence of density fields, each of which is decompressed into a weighted sum of a set of radial basis functions (RBFs) and an optional residual field. Source radiances from single and optionally multiple scattering are directly computed at only the RBF centers and then approximated at other points in the volume using an RBF-based interpolation. Using the computed source radiances, a ray marching technique using slice-based integration of radiance along each viewing ray is performed to render the final image. During the ray marching process, the residual field may be compensated back into the radiance integral to generate images of higher detail. | 01-01-2009 |

20090006046 | Real-Time Rendering of Light-Scattering Media - A real-time algorithm for rendering of an inhomogeneous scattering media such as smoke under dynamic low-frequency environment lighting is described. An input media animation is represented as a sequence of density fields, each of which is represented by an approximate model density field and a residual density field. The algorithm uses the approximate model density field to compute an approximate source radiance, and further computes an effective exitant radiance by compositing the approximate source radiance using a compositing methods such as ray marching. During the compositing process (e.g., ray marching), the residual field is compensated back into the radiance integral to generate images of higher detail. | 01-01-2009 |

20090006047 | Real-Time Rendering of Light-Scattering Media - A real-time algorithm for rendering of an inhomogeneous scattering media such as smoke under dynamic low-frequency environment lighting is described. An input media animation is represented as a sequence of density fields, each of which is decomposed into a weighted sum of a set of radial basis functions (RBFs) and an optional residual field. Source radiances from single and optionally multiple scattering are directly computed at only the RBF centers and then approximated at other points in the volume using an RBF-based interpolation. Unique approximation techniques are introduced in the computational algorithms to simplify and speed up the computation of source radiance contributed by single and multiple scattering. Using the computed source radiances, a ray marching technique using slice-based integration of radiance along each viewing ray may be performed to render the final image. | 01-01-2009 |

20090006051 | Real-Time Rendering of Light-Scattering Media - A real-time algorithm for rendering an inhomogeneous scattering medium such as fog is described. An input media animation is represented as a sequence of density fields, each of which is decomposed into a weighted sum of a set of radial basis functions (RBFs) such as Gaussians. The algorithm computes airlight and surface reflectance of the inhomogeneous scattering medium. Several approximations are taken which lead to analytical solutions of quantities such as an optical depth integrations and single scattering integrations, and a reduced number of integrations that need to be calculated. The resultant algorithm is able to render inhomogeneous media including their shadowing and scattering effects in real time. The algorithm may be adopted for a variety of light sources including point lights and environmental lights. | 01-01-2009 |

20090006052 | Real-Time Rendering of Light-Scattering Media - A real-time algorithm for rendering of an inhomogeneous scattering medium such as fog with a surface object immersed therein is described. An input media animation is represented as a sequence of density fields. The algorithm computes surface reflectance of the surface object in the inhomogeneous scattering medium. The algorithm may also compute airlight of the inhomogeneous scattering medium. Several approximations are taken which lead to analytical solutions of quantities such as optical depth integrations and single scattering integrations, and a reduced number of integrations that need to be calculated. The resultant algorithm is able to render inhomogeneous media including their shadowing and scattering effects in the real time. The algorithm may be adopted for a variety of light sources including point lights and environmental lights. | 01-01-2009 |

20090022414 | HIGH DYNAMIC RANGE IMAGE HALLUCINATION - An apparatus and method provide for providing an output image from an input image. The input image may contain at least one portion that does not display certain desired information of the image, such as texture information. The desired information may be obtained from a second portion of the input image and applied to the at least one portion that does not contain the texture information or contains a diminished amount of the texture information. Also, at least one characteristic of the second portion of the input image may not be applied to the at least one portion such as illumination information. In another example, the input image may be decomposed into multiple parts such as a high frequency and a low frequency component. Each component may be hallucinated individually or independently and combined to form the output image. | 01-22-2009 |

20090244083 | INVERSE TEXTURE SYNTHESIS - A “texture generator” uses an inverse texture synthesis solution that runs in the opposite direction to traditional forward synthesis techniques to construct 2D texture compactions for use by a graphics processing unit (GPU) of a computer system. These small 2D texture compactions generally summarize an original globally variant texture or image, and are used to reconstruct the original texture or image, or to re-synthesize new textures or images under user-supplied constraints. In various embodiments, the texture generator uses the texture compaction to provide real-time synthesis of globally variant textures on a GPU, where texture memory is generally too small for large textures. Further, the texture generator provides an optimization framework for inverse texture synthesis which ensures that each input region is properly encoded in the output compaction. In addition, the texture generator also computes orientation fields for anisotropic textures containing both low- and high-frequency regions. | 10-01-2009 |

20090322769 | BULK-SYNCHRONOUS GRAPHICS PROCESSING UNIT PROGRAMMING - Described is a technology in a computing environment comprising a programming language for general purpose computation on a graphics processing unit (GPU), along with an associated compiler. A Bulk-Synchronous GPU Programming (BSGP) program is programmed to include barriers to describe parallel processing on GPUs. A BSGP compiler detects barriers corresponding to supersteps, converts BSGP programs to kernels based on the barriers, and combines them. During compilation, the compiler aligns barriers in the statements and bundles the corresponding supersteps together. A par construct is provided to allow the programmer to control aspects of bundling, e.g., by specifying a block independent statements. Thread manipulation emulation is provided to transparently emulate thread creation and destruction, with operations fork and kill. Also provided is remote variable access intrinsics for efficient communications between threads, and collective primitive operations. | 12-31-2009 |

20100033482 | Interactive Relighting of Dynamic Refractive Objects - Dynamic refractive object relighting technique embodiments are presented which involve rendering an image of a refractive object in a dynamic scene by first voxelizing a representation of the surfaces of the object into a volumetric representation in the form of a rectangular voxel grid. A refractive index is assigned to each voxel based on user-input material parameters. Next, the paths of photons are traced in a step-wise manner as each photon refracts through the object. The size of each step forward is variable and based on variations in refractive index of the object. Radiance values are assigned to all the voxels that the photons traverse in their paths through the object. An output image of the refractive object is then rendered from a user-input viewpoint by tracing viewing rays from the viewpoint into the scene and calculating the amount of radiance that reaches the viewpoint along each of the rays. | 02-11-2010 |

20100033488 | Example-Based Motion Detail Enrichment in Real-Time - An approach to enrich skeleton-driven animations with physically-based secondary deformation in real time is described. To achieve this goal, the technique described employs a surface-based deformable model that can interactively emulate the dynamics of both low- and high-frequency volumetric effects. Given a surface mesh and a few sample sequences of its physical behavior, a set of motion parameters of the material are learned during an off-line preprocessing step. The deformable model is then applicable to any given skeleton-driven animation of the surface mesh. Additionally, the described dynamic skinning technique can be entirely implemented on GPUs and executed with great efficiency. Thus, with minimal changes to the conventional graphics pipeline, the technique can drastically enhance the visual experience of skeleton-driven animations by adding secondary deformation in real time. | 02-11-2010 |

20100079451 | RAY TRACING ON GRAPHICS HARDWARE USING KD-TREES - Described is a technology by which a ray tracer incorporates a GPU-based kd-tree builder for rendering arbitrary dynamic scenes. For each frame, the ray tracer builds a kd-tree for the scene geometry. The ray tracer spawns and traces eye rays, reflective and refractive rays, and shadow rays. For each ray to be traced, the ray tracer walks through the kd-tree until it reaches leaf nodes and associated triangles. When a ray passes through both sides of a splitting plane, the “far” sub-tree is pushed into the stack and the “near” sub-tree is traversed first. | 04-01-2010 |

20100079452 | PHOTON MAPPING ON GRAPHICS HARDWARE USING KD-TREES - Described is a technology by which a GPU-based photon mapping mechanism/algorithm uses a kd-tree to render arbitrary dynamic scenes. For each frame, the mechanism emits and traces a set of photons into the scene. When a photon hits a surface, it can either be reflected, transmitted, or absorbed based on the surface material. Once photon tracing is done, a kd-tree is built for the stored photons. To estimate the radiance value at an arbitrary surface point, the k-nearest photons are located and filtered. The photon tracing and photon kd-tree construction, as well as the radiance estimation using k-nearest neighbor (KNN) searches are performed on graphics hardware, e.g., a GPU. In one example, only caustic photons are traced, whereby a photon is terminated and stored once it hits a diffuse surface. | 04-01-2010 |

20100082703 | OCTREE CONSTRUCTION ON GRAPHICS PROCESSING UNITS - An octree GPU construction system and method for constructing a complete octree data structure on a graphics processing unit (GPU). Embodiments of the octree GPU construction system and method first defines a complete octree data structure as forming a complete partition of the 3-D space and including a vertex, edge, face, and node arrays, and neighborhood information. Embodiments of the octree GPU construction system and method input a point cloud and construct a node array. Next, neighboring nodes are computed for each of the nodes in the node arrays by using at least two pre-computed look-up tables (such as a parent look-up table and a child look-up table). Embodiments of the octree GPU construction system and method then use the neighboring nodes and neighborhood information to compute a vertex array, edge array, and face array are computed by determining owner information and self-ownership information based on the neighboring nodes. | 04-01-2010 |

20100082704 | REAL-TIME KD-TREE CONSTRUCTION ON GRAPHICS HARDWARE - Described is a technology for constructing kd-trees on GPUs, in a manner that is sufficiently fast to achieve real-time performance by exploiting GPU-bsaed parallelism during the kd-tree construction. Tree nodes are built in breadth-first search order, e.g., to use a thread for each node at each level. For large nodes at upper tree levels, computations are parallelized over geometric primitives (instead of nodes). To this end, large nodes are split into child nodes by cutting off empty space based until an empty space ratio is achieved, and thereafter performing spatial splitting. Small nodes are split based on split candidate costs, e.g., computed by a surface area heuristic or a voxel volume heuristic (VVH). | 04-01-2010 |

20100085352 | PARALLEL SURFACE RECONSTRUCTION - Described is a technology in which point cloud surface reconstruction is performed via parallel processing on a graphics processing unit, achieving real-time reconstruction rates. An octree is built for a given set of oriented points, with each node containing a set of points enclosed by the node. The data structure is built on the GPU, in parallel, using level-order traversals to process nodes at a same tree level. The surface is reconstructed based on data configured and located via the traversals. To produce the surface, an implicit function over the volume spanned by the octree nodes is computed using the GPU, e.g., based on a Poisson surface reconstruction method. A sparse linear system is built and a multi-grid solver is employed to solve the system. An adaptive marching cubes procedure is performed on the GPU to extract an isosurface of the implicit function as a triangular mesh | 04-08-2010 |

20100085353 | USER-GUIDED SURFACE RECONSTRUCTION - Described is a technology by which a user interacts with a surface representative of a point cloud data to correct for imperfect scan data. The surface is reconstructed based on the interaction. Real time viewing of the image is facilitated by parallel surface reconstruction. For example, the user may draw strokes to reduce topological ambiguities in poorly-sampled areas. An algorithm automatically adds new oriented sample points to the original point cloud based on the user interaction. Then a new isosurface is generated for the augmented point cloud. The user also may specify the geometry of missing areas of the surface. The user copies a set of points from another point cloud, and places the points around the target area. A new isosurface is then generated. | 04-08-2010 |

20100085360 | RENDERING IN SCATTERING MEDIA - Techniques are described for rendering a volume of scattering media, in particular by computing radiances of points or voxels in the scattering media. A set of sample points in the scattering media are found. Radiances of the sample points are computed. Radiance gradients of the sample points are computed from the radiances. The radiances and gradients are used to interpolate radiances throughout the scattering media. The set of sample points may be computed in an iterative dynamic manner in order to concentrate samples near features (e.g., shadow edges) of the scattering media. | 04-08-2010 |

20110208492 | Joint-Aware Manipulation of Deformable Models - This disclosure describes a joint-aware deformation framework that supports the direct manipulation of an arbitrary mix of rigid and deformable components. The deformation framework may include at least a joint-analysis and a joint-aware deformation enabling a more realistic deformation of a joint-aware model. | 08-25-2011 |