US9767598B2 - Smoothing and robust normal estimation for 3D point clouds - Google Patents
Smoothing and robust normal estimation for 3D point clouds Download PDFInfo
- Publication number
- US9767598B2 US9767598B2 US13/566,796 US201213566796A US9767598B2 US 9767598 B2 US9767598 B2 US 9767598B2 US 201213566796 A US201213566796 A US 201213566796A US 9767598 B2 US9767598 B2 US 9767598B2
- Authority
- US
- United States
- Prior art keywords
- point
- output
- points
- input
- plane
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/04—Texture mapping
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/05—Geographic models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/08—Volume rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
- G06T15/205—Image-based rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- H04N13/0014—
-
- H04N13/0239—
-
- H04N13/0257—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
- H04N13/117—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/194—Transmission of image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/243—Image signal generators using stereoscopic image cameras using three or more 2D image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/246—Calibration of cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/257—Colour aspects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/142—Constructional details of the terminal equipment, e.g. arrangements of the camera and the display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
- H04N7/157—Conference systems defining a virtual conference space and using avatars or agents
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/56—Particle system, point based geometry or rendering
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2227/00—Details of public address [PA] systems covered by H04R27/00 but not provided for in any of its subgroups
- H04R2227/005—Audio distribution systems for home, i.e. multi-room use
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/15—Aspects of sound capture and related signal processing for recording or reproduction
Definitions
- 3D scanning tools are often used to scan a 3D surface to generate corresponding 3D point clouds. These point clouds are then typically used for constructing 3D point- or mesh-based digital models of the scanned surface. Unfortunately, due to many possible sources of noise during the scanning process, the resulting 3D models tend to be noisy.
- Techniques such as mesh smoothing operate by filtering or otherwise adjusting a 3D input surface to increase a degree of smoothness of that surface by denoising the data representing that surface.
- one recent technique provides a bilateral denoising filter for 3D point clouds that operates by filtering vertices of a corresponding 3D mesh. This technique generally filters vertices in the normal direction using local neighborhoods to denoise the mesh while partially preserving local features.
- existing techniques for smoothing or denoising 3D surfaces, models, meshes or point clouds tend to remove noise while partially blurring out fine details of 3D features. Further, these methods tend to produce at least some degree of shrinkage or drifting of the 3D surfaces.
- a “Point Cloud Smoother,” as described herein, provides various techniques for refining a 3D point cloud or other 3D input model to generate a smoothed and denoised 3D output model by robustly fitting planes to each point of the input model and using those planes to estimate new points and corresponding normals of the 3D output model.
- These techniques are useful for a number of purposes, including, but not limited to, generation of free viewpoint video (FVV) which allows 3D data of videos or images to be rendered and viewed from any desired viewpoint that is supported by the input data.
- FVV free viewpoint video
- the 3D smoothing techniques enabled by the Point Cloud Smoother generally begin by fitting small planes to localized regions of a 3D input model or point cloud using a robust estimator, and then using these planes to identify new points corresponding to a 3D output model.
- plane fitting is performed for each point in the input model by first finding a set of the nearest neighbors for each point.
- Various robust estimation techniques e.g., RANSAC, MLESAC, LMS, MUSE, ALKS, RESC, ASSC, etc.
- Each point is then projected some or all of the way along a corresponding normal of the corresponding plane to the surface of that plane. Together, the projected points of the input model represent a new set of output points corresponding to a 3D output model.
- a new normal for each point of the 3D output model is determined by finding a set of the nearest neighbors to each output point, fitting a new plane to each output point and its set of nearest neighbors, and computing a normal direction for each plane. The normal direction for each of these new planes is then assigned to each corresponding point to complete the 3D output model.
- FIG. 1 provides an exemplary architectural flow diagram that illustrates program modules for implementing various embodiments of the Point Cloud Smoother, as described herein.
- FIG. 2 illustrates a general system flow diagram that illustrates exemplary methods for implementing various embodiments of the Point Cloud Smoother, as described herein.
- FIG. 3 is a general system diagram depicting a simplified general-purpose computing device having simplified computing and I/O capabilities for use in implementing various embodiments of the Point Cloud Smoother, as described herein.
- a “Point Cloud Smoother,” as described herein, provides various techniques for refining a 3D point clouds or other 3D input models to generate smoothed and denoised 3D output models. Smoothing and denoising is achieved, in part, by robustly fitting planes to a neighborhood of points around and including each point of the input model and using those planes to estimate new points of the 3D output model. Normals for the points of the 3D output model are then determined by performing another round of plane fitting to a neighborhood of points around and including each point of the 3D output model. Normal directions or vectors are then computed for each of these planes and assigned to each corresponding point of the 3D output model.
- FVV free viewpoint video
- Point Cloud Smoother allows 3D data of videos or images to be denoised and then rendered and viewed from any desired viewpoint that is supported by the input data.
- the “Point Cloud Smoother,” provides various techniques for refining a 3D point clouds or other 3D input models to generate smoothed and denoised 3D output models.
- the processes summarized above are illustrated by the general system diagram of FIG. 1 .
- the system diagram of FIG. 1 illustrates the interrelationships between program modules for implementing various embodiments of the Point Cloud Smoother, as described herein.
- the system diagram of FIG. 1 illustrates a high-level view of various embodiments of the Point Cloud Smoother
- FIG. 1 is not intended to provide an exhaustive or complete illustration of every possible embodiment of the Point Cloud Smoother as described throughout this document.
- any boxes and interconnections between boxes that may be represented by broken or dashed lines in FIG. 1 represent alternate embodiments of the Point Cloud Smoother described herein, and that any or all of these alternate embodiments, as described below, may be used in combination with other alternate embodiments that are described throughout this document.
- the processes enabled by the Point Cloud Smoother begin operation by using a data input module 100 to receive a set of 3D input points 110 representing a 3D point cloud, 3D mesh, or other point-based 3D model or object.
- the 3D input points 110 are derived from pre-existing 3D models or point clouds, or 3D point clouds generated by any desired 3D scanning technique (e.g., stereo depth evaluation techniques, laser depth scanning, etc.).
- a neighborhood ID module 120 is then used to evaluate each point in the set of 3D input points 110 , to identify a set of nearest neighbors around each input point.
- the number of neighbors is fixed at some relatively small number or is optionally set or adjusted to any desired value via a user input module 130 or other input mechanism.
- computational time for identifying the nearest neighbors of each input point is reduced by pre-processing the set of 3D input points 110 to construct a corresponding 3D index tree of the input points from which neighboring points can be quickly selected or identified.
- a plane fitting module 140 is used to fit a plane in 3D space to each input point and its corresponding set of nearest neighbors using a robust estimator (e.g., RANSAC, MLESAC, LMS, MUSE, ALKS, RESC, ASSC, etc.) or other data or shape fitting technique.
- a projection module 150 then uses these planes to create a set of 3D output points 160 by projecting each input point onto its corresponding plane. The resulting intersection of the 3D input point with its corresponding plane represents the 3D spatial location of each corresponding 3D output point.
- the set of 3D output points 160 is then passed back to the neighborhood IS module 120 that evaluates each point in the set of 3D output points 160 , to identify a set of nearest neighbors around each output point.
- the number of neighbors around each output point is fixed at some relatively small number or is optionally set or adjusted to any desired value via the user input module 130 or other input mechanism. Note that the number of neighbors around input points and output points can be the same, if desired, though there is no requirement for the number of neighbors to be the same.
- computational time for identifying the nearest neighbors of each output point is optionally reduced by pre-processing the set of 3D output points 160 to construct a corresponding 3D index tree of the output points from which neighboring points can be quickly selected or identified.
- the plane fitting module 140 is used to fit a plane in 3D space to each output point and its corresponding set of nearest neighbors using a robust estimator or other data or shape fitting technique.
- the planes associated with the set of 3D output points 160 are then passed to a surface normal module 170 that uses each of these planes to compute corresponding normal directions or vectors for each plane.
- Each computed normal is then assigned to the corresponding output point to generate a smoothed and denoised 3D output model 180 .
- One or more further iterations of smoothing are then optionally performed by providing the 3D output model 180 back to the data input module 100 for use as a new set of 3D input points 110 that is then used to generate a new 3D output model using the processes and techniques described above.
- a normal correction module 190 is used to evaluate normal directions for points in localized regions of the 3D output model relative to the normal direction of neighboring points. It has been observed that where the input data is highly noisy, in rare cases, normals may occasionally be inverted or reversed (i.e., 180 degrees in the wrong direction) relative to surrounding points. In such cases, the normal correction module 190 acts to flip or reverse the normal direction of one or more points to correspond to the general direction of the normal direction of those neighbors. Note that the number of neighbors evaluated for this purpose can be set at a predetermined value, or set or adjusted to any number desired.
- the Point Cloud Smoother provides various techniques for refining a 3D point clouds or other 3D input models to generate smoothed and denoised 3D output models.
- the following sections provide a detailed discussion of the operation of various embodiments of the Point Cloud Smoother, and of exemplary methods for implementing the program modules described in Section 1 with respect to FIG. 1 .
- a “Point Cloud Smoother,” as described herein, provides various techniques for refining a 3D point cloud or other 3D input model to generate a smoothed and denoised 3D output model by robustly fitting planes to each point of the input model and using those planes to estimate new points and corresponding normals of the 3D output model.
- These techniques are useful for a number of purposes, including, but not limited to, generation of free viewpoint video (FVV) which allows 3D data of videos or images to be rendered and viewed from any desired viewpoint that is supported by the input data.
- FVV free viewpoint video
- a 3D point cloud is a set of points or vertices in a 3D coordinate system. These vertices are usually defined by x, y, and z coordinates, and typically represent the external surface of an object.
- 3D point clouds or models can be generated or constructed using a wide variety of techniques. For example, Point clouds are often created by 3D scanners, including laser-based scanners, LIDAR systems, etc., and may also be created using other techniques such as, for example, stereo imaging where multiple images of a scene are used to construct pixel or point-based depth maps of scenes or objects in a scene. In general, 3D scanners process objects to identify large numbers of surface points on the object to produce a 3D point cloud representing the object surface.
- mesh vertices can be treated as individual points to provide the point cloud that is processed by the Point Cloud Smoother.
- 3D point clouds or models used as input can be pre-filtered or pre-processed in any manner desired (e.g., remove or attenuate outliers) prior to processing by the Point Cloud Smoother, though such pre-processing is generally not necessary.
- a number of robust estimation techniques have been developed for fitting data. These techniques include, but are not limited to, random sample consensus (RANSAC), maximum-likelihood sample consensus (MLESAC), least median of squares (LMS), minimum unbiased scale estimator (MUSE), adaptive least k th order squares (ALKS), residual sample consensus (RESC), adaptive scale sample consensus (ASSC), etc.
- RANSAC random sample consensus
- MLESAC maximum-likelihood sample consensus
- LMS least median of squares
- MUSE minimum unbiased scale estimator
- ALKS adaptive least k th order squares
- RESC residual sample consensus
- ASSC adaptive scale sample consensus
- any desired data or shape fitting technique can be adapted for use with the Point Cloud Smoother to fit planes to sets of 3D data points (i.e., each point and its set of neighbors).
- the planes resulting from the data fitting process are used as a surface onto which the 3D input points are projected along the normal of each point.
- the resulting intersection of the 3D input point with its corresponding plane represents the 3D spatial location of each corresponding 3D output point.
- the planes resulting from the data fitting process are used to compute normal vectors or directions (i.e., perpendicular to the plane), with the resulting vector or direction being assigned to the corresponding 3D output point.
- the combination of 3D output points and corresponding normals then defines a 3D output point cloud that can be used directly, or converted to any desired type of 3D output model.
- Point Cloud Smoother Given the smoothed and denoised 3D point cloud and associated normal produced as output by the Point Cloud Smoother, many well-known techniques exist for converting that point cloud to any of a number of different formats or model types for use in a wide range of 3D modeling applications. For example, while point clouds can be directly rendered, point clouds themselves are generally not directly usable in most 3D applications, and therefore are usually converted to polygon or triangle mesh models, NURBS surface models, or CAD models through a process commonly referred to as surface reconstruction. There are many existing techniques for converting a point cloud to a 3D surface.
- Point Cloud Smoother is adaptable for use with any such techniques to construct 3D output models of any desired type.
- FIG. 2 provides an exemplary operational flow diagram that summarizes the operation of some of the various embodiments of the Point Cloud Smoother. Note that FIG. 2 is not intended to be an exhaustive representation of all of the various embodiments of the Point Cloud Smoother described herein, and that the embodiments represented in FIG. 2 are provided only for purposes of explanation.
- any boxes and interconnections between boxes that are represented by broken or dashed lines in FIG. 2 represent optional or alternate embodiments of the Point Cloud Smoother described herein, and that any or all of these optional or alternate embodiments, as described below, may be used in combination with other alternate embodiments that are described throughout this document.
- the Point Cloud Smoother begins operation by receiving 200 a set of 3D input points representing a 3D point cloud, 3D mesh, or other point-based 3D model or object.
- pre-existing models or point clouds, or 3D point clouds generated by any desired 3D scanning technique may be used as input by the Point Cloud Smoother to generate a smoothed 3D output model.
- the Point Cloud Smoother uses those 3D input points to determine 210 or identify a set of the nearest j neighbors of each input point.
- the number of neighbors is optionally set or adjusted via a user interface.
- computational time for identifying nearest neighbors of each input point is reduced by pre-processing the set of 3D input points to construct a corresponding 3D index tree of the input points from which neighboring points can be quickly selected.
- the Point Cloud Smoother fits 220 a plane in 3D space through each input point and its corresponding set of nearest neighbors using a robust estimator (e.g., RANSAC, MLESAC, LMS, MUSE, ALKS, RESC, ASSC, etc.).
- the Point Cloud Smoother uses these planes to create a set of 3D output points by projecting 230 each input point onto its corresponding plane.
- the resulting intersection of the 3D input point with its corresponding plane represents the 3D spatial location of each corresponding 3D output point.
- the Point Cloud Smoother uses the set of 3D output points to determine 240 or identify a set of the nearest k neighbors of each input point.
- the number of neighbors, k, for each output point may be the same as the number of neighbors, j, used for each input point, there is no requirement that j and k are the same.
- the number of neighbors is optionally set or adjusted via a user interface.
- computational time for identifying nearest neighbors of each output point is reduced by pre-processing the set of 3D output points to construct a corresponding 3D index tree of the output points from which neighboring points can be quickly selected.
- the Point Cloud Smoother fits 250 a plane in 3D space through each output point and its corresponding set of nearest neighbors using a robust estimator.
- the Point Cloud Smoother uses each of these planes to compute 260 corresponding normal vectors for each output point.
- the Point Cloud Smoother provides 270 a smoothed 3D output model 180 as the combination of the 3D output points and the corresponding surface normals.
- One or more further iterations of smoothing are then optionally performed by providing the 3D output model 180 to the Point Cloud Smoother for use as a new set of 3D input points that is then used to generate a new 3D output model using the processes and techniques described above. Note that the number of neighboring points around both input points and output points can be adjusted to any desired values for any additional iterations of smoothing.
- FIG. 3 illustrates a simplified example of a general-purpose computer system on which various embodiments and elements of the Point Cloud Smoother, as described herein, may be implemented. It should be noted that any boxes that are represented by broken or dashed lines in FIG. 3 represent alternate embodiments of the simplified computing device, and that any or all of these alternate embodiments, as described below, may be used in combination with other alternate embodiments that are described throughout this document.
- FIG. 3 shows a general system diagram showing a simplified computing device such as computer 300 .
- Such computing devices can be typically be found in devices having at least some minimum computational capability, including, but not limited to, personal computers, server computers, hand-held computing devices, laptop or mobile computers, communications devices such as cell phones and PDA's, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, audio or video media players, etc.
- the device should have a sufficient computational capability and system memory to enable basic computational operations.
- the computational capability is generally illustrated by one or more processing unit(s) 310 , and may also include one or more GPUs 315 , either or both in communication with system memory 320 .
- the processing unit(s) 310 of the general computing device of may be specialized microprocessors, such as a DSP, a VLIW, or other micro-controller, or can be conventional CPUs having one or more processing cores, including specialized GPU-based cores in a multi-core CPU.
- the simplified computing device of FIG. 3 may also include other components, such as, for example, a communications interface 330 .
- the simplified computing device of FIG. 3 may also include one or more conventional computer input devices 340 (e.g., pointing devices, keyboards, audio input devices, video input devices, haptic input devices, devices for receiving wired or wireless data transmissions, etc.).
- the simplified computing device of FIG. 3 may also include other optional components, such as, for example, one or more conventional computer output devices 350 (e.g., display device(s) 355 , audio output devices, video output devices, devices for transmitting wired or wireless data transmissions, etc.).
- typical communications interfaces 330 , input devices 340 , output devices 350 , and storage devices 360 for general-purpose computers are well known to those skilled in the art, and will not be described in detail herein.
- the simplified computing device of FIG. 3 may also include a variety of computer readable media.
- Computer readable media can be any available media that can be accessed by computer 300 via storage devices 360 and includes both volatile and nonvolatile media that is either removable 370 and/or non-removable 380 , for storage of information such as computer-readable or computer-executable instructions, data structures, program modules, or other data.
- Computer readable media may comprise computer storage media and communication media.
- Computer storage media includes, but is not limited to, computer or machine readable media or storage devices such as DVD's, CD's, floppy disks, tape drives, hard drives, optical drives, solid state memory devices, RAM, ROM, EEPROM, flash memory or other memory technology, magnetic cassettes, magnetic tapes, magnetic disk storage, or other magnetic storage devices, or any other device which can be used to store the desired information and which can be accessed by one or more computing devices.
- computer or machine readable media or storage devices such as DVD's, CD's, floppy disks, tape drives, hard drives, optical drives, solid state memory devices, RAM, ROM, EEPROM, flash memory or other memory technology, magnetic cassettes, magnetic tapes, magnetic disk storage, or other magnetic storage devices, or any other device which can be used to store the desired information and which can be accessed by one or more computing devices.
- modulated data signal or “carrier wave” generally refer a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
- communication media includes wired media such as a wired network or direct-wired connection carrying one or more modulated data signals, and wireless media such as acoustic, RF, infrared, laser, and other wireless media for transmitting and/or receiving one or more modulated data signals or carrier waves. Combinations of the any of the above should also be included within the scope of communication media.
- Point Cloud Smoother software, programs, and/or computer program products embodying the some or all of the various embodiments of the Point Cloud Smoother described herein, or portions thereof, may be stored, received, transmitted, or read from any desired combination of computer or machine readable media or storage devices and communication media in the form of computer executable instructions or other data structures.
- Point Cloud Smoother described herein may be further described in the general context of computer-executable instructions, such as program modules, being executed by a computing device.
- program modules include routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types.
- the embodiments described herein may also be practiced in distributed computing environments where tasks are performed by one or more remote processing devices, or within a cloud of one or more devices, that are linked through one or more communications networks.
- program modules may be located in both local and remote computer storage media including media storage devices.
- the aforementioned instructions may be implemented, in part or in whole, as hardware logic circuits, which may or may not include a processor.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Graphics (AREA)
- General Physics & Mathematics (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Remote Sensing (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
-
- an operational overview of the Point Cloud Smoother;
- 3D point clouds and models;
- fitting planes using robust estimators or other data fitting techniques;
- projecting 3D points onto fitted planes; and
- computing of surface normals for points of the 3D output model.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/566,796 US9767598B2 (en) | 2012-05-31 | 2012-08-03 | Smoothing and robust normal estimation for 3D point clouds |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261653983P | 2012-05-31 | 2012-05-31 | |
US13/566,796 US9767598B2 (en) | 2012-05-31 | 2012-08-03 | Smoothing and robust normal estimation for 3D point clouds |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130321393A1 US20130321393A1 (en) | 2013-12-05 |
US9767598B2 true US9767598B2 (en) | 2017-09-19 |
Family
ID=60922749
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/566,796 Active 2035-02-23 US9767598B2 (en) | 2012-05-31 | 2012-08-03 | Smoothing and robust normal estimation for 3D point clouds |
Country Status (1)
Country | Link |
---|---|
US (1) | US9767598B2 (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160292829A1 (en) * | 2012-06-25 | 2016-10-06 | Yoldas Askan | Method of generating a smooth image from point cloud data |
US20180025496A1 (en) * | 2016-07-19 | 2018-01-25 | Qualcomm Incorporated | Systems and methods for improved surface normal estimation |
US20190043250A1 (en) * | 2012-06-25 | 2019-02-07 | Yoldas Askan | Method of generating a smooth image from point cloud data |
CN110458780A (en) * | 2019-08-14 | 2019-11-15 | 上海眼控科技股份有限公司 | 3D point cloud data de-noising method, apparatus, computer equipment and readable storage medium storing program for executing |
CN110719497A (en) * | 2018-07-12 | 2020-01-21 | 华为技术有限公司 | Point cloud coding and decoding method and coder-decoder |
US10648832B2 (en) * | 2017-09-27 | 2020-05-12 | Toyota Research Institute, Inc. | System and method for in-vehicle display with integrated object detection |
US20220028119A1 (en) * | 2018-12-13 | 2022-01-27 | Samsung Electronics Co., Ltd. | Method, device, and computer-readable recording medium for compressing 3d mesh content |
US11321862B2 (en) | 2020-09-15 | 2022-05-03 | Toyota Research Institute, Inc. | Systems and methods for multi-camera modeling with neural camera networks |
US11494927B2 (en) | 2020-09-15 | 2022-11-08 | Toyota Research Institute, Inc. | Systems and methods for self-supervised depth estimation |
US11508080B2 (en) | 2020-09-15 | 2022-11-22 | Toyota Research Institute, Inc. | Systems and methods for generic visual odometry using learned features via neural camera models |
US11615544B2 (en) | 2020-09-15 | 2023-03-28 | Toyota Research Institute, Inc. | Systems and methods for end-to-end map building from a video sequence using neural camera models |
US11995895B2 (en) * | 2019-06-03 | 2024-05-28 | Nvidia Corporation | Multi-object tracking using correlation filters in video analytics applications |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9002719B2 (en) | 2012-10-08 | 2015-04-07 | State Farm Mutual Automobile Insurance Company | Device and method for building claim assessment |
US10049281B2 (en) * | 2012-11-12 | 2018-08-14 | Shopperception, Inc. | Methods and systems for measuring human interaction |
US9530225B1 (en) * | 2013-03-11 | 2016-12-27 | Exelis, Inc. | Point cloud data processing for scalable compression |
US9082015B2 (en) | 2013-03-15 | 2015-07-14 | State Farm Mutual Automobile Insurance Company | Automatic building assessment |
US8818572B1 (en) | 2013-03-15 | 2014-08-26 | State Farm Mutual Automobile Insurance Company | System and method for controlling a remote aerial device for up-close inspection |
US8756085B1 (en) * | 2013-03-15 | 2014-06-17 | State Farm Mutual Automobile Insurance Company | Systems and methods for assessing property damage |
US8872818B2 (en) | 2013-03-15 | 2014-10-28 | State Farm Mutual Automobile Insurance Company | Methods and systems for capturing the condition of a physical structure |
CN104424655A (en) * | 2013-09-10 | 2015-03-18 | 鸿富锦精密工业(深圳)有限公司 | System and method for reconstructing point cloud curved surface |
CN104050709B (en) * | 2014-06-06 | 2017-08-29 | 联想(北京)有限公司 | A kind of three dimensional image processing method and electronic equipment |
US10750153B2 (en) | 2014-09-22 | 2020-08-18 | Samsung Electronics Company, Ltd. | Camera system for three-dimensional video |
US11205305B2 (en) | 2014-09-22 | 2021-12-21 | Samsung Electronics Company, Ltd. | Presentation of three-dimensional video |
CN106033620B (en) * | 2015-03-13 | 2018-10-19 | 腾讯科技(深圳)有限公司 | A kind of point cloud model restorative procedure, device and computing device |
US9582939B2 (en) | 2015-06-11 | 2017-02-28 | Nokia Technologies Oy | Structure preserved point cloud simplification |
US9934590B1 (en) | 2015-06-25 | 2018-04-03 | The United States Of America As Represented By The Secretary Of The Air Force | Tchebichef moment shape descriptor for partial point cloud characterization |
CN105096268B (en) * | 2015-07-13 | 2018-02-02 | 西北农林科技大学 | One kind point cloud denoising smooth method |
US10066346B2 (en) * | 2015-08-12 | 2018-09-04 | Topcon Positioning Systems, Inc. | Point cloud based surface construction |
RU2612571C1 (en) * | 2015-11-13 | 2017-03-09 | Общество с ограниченной ответственностью "ХЕЛЬГИ ЛАБ" | Method and system for recognizing urban facilities |
US10176527B1 (en) | 2016-04-27 | 2019-01-08 | State Farm Mutual Automobile Insurance Company | Providing shade for optical detection of structural features |
CN105844600B (en) * | 2016-04-27 | 2018-03-16 | 北京航空航天大学 | A kind of extraterrestrial target three-dimensional point cloud fairing denoising method |
CN106372283B (en) * | 2016-08-24 | 2018-06-08 | 大连理工大学 | A kind of thin wall obtained towards digital photography surveys three-dimensional appearance Processing Method of Point-clouds |
US10837773B2 (en) | 2016-12-30 | 2020-11-17 | DeepMap Inc. | Detection of vertical structures based on LiDAR scanner data for high-definition maps for autonomous vehicles |
US11665308B2 (en) | 2017-01-31 | 2023-05-30 | Tetavi, Ltd. | System and method for rendering free viewpoint video for sport applications |
RU2638638C1 (en) * | 2017-02-14 | 2017-12-14 | Общество с ограниченной ответственностью "Хельги Лаб" (ООО "Хельги Лаб") | Method and system of automatic constructing three-dimensional models of cities |
KR102320198B1 (en) | 2017-04-05 | 2021-11-02 | 삼성전자주식회사 | Method and apparatus for refining depth image |
CN107492072A (en) * | 2017-07-05 | 2017-12-19 | 山东理工大学 | Dispersion point cloud normal estimation method based on sampling point neighborhood isomorphism curved surface |
US11049218B2 (en) | 2017-08-11 | 2021-06-29 | Samsung Electronics Company, Ltd. | Seamless image stitching |
US10684537B2 (en) * | 2017-11-14 | 2020-06-16 | Texas Instruments Incorporated | Camera-assisted arbitrary surface characterization and correction |
CN111788602B (en) * | 2017-12-29 | 2024-05-28 | 泰立戴恩菲力尔有限责任公司 | Point cloud denoising system and method |
US11250594B2 (en) * | 2019-01-09 | 2022-02-15 | Tencent America LLC | Method and apparatus for geometry smoothing by local geometry projection |
US10904579B2 (en) * | 2019-01-09 | 2021-01-26 | Tencent America LLC | Method and apparatus for annealing iterative geometry smoothing |
CN110223382B (en) * | 2019-06-13 | 2021-02-12 | 电子科技大学 | Single-frame image free viewpoint three-dimensional model reconstruction method based on deep learning |
WO2021103013A1 (en) * | 2019-11-29 | 2021-06-03 | 深圳市大疆创新科技有限公司 | Methods for data encoding and data decoding, device, and storage medium |
US11315299B1 (en) | 2020-11-13 | 2022-04-26 | Unity Technologies Sf | Method for computation of local densities for virtual fibers |
US20220300681A1 (en) * | 2021-03-16 | 2022-09-22 | Yuan Ren | Devices, systems, methods, and media for point cloud data augmentation using model injection |
WO2023212575A1 (en) * | 2022-04-25 | 2023-11-02 | Virginia Tech Intellectual Properties, Inc. | Automated objects labeling in video data for machine learning and other classifiers |
WO2024174092A1 (en) * | 2023-02-21 | 2024-08-29 | Oppo广东移动通信有限公司 | Encoding/decoding method, code stream, encoder, decoder, and storage medium |
Citations (92)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5850352A (en) | 1995-03-31 | 1998-12-15 | The Regents Of The University Of California | Immersive video, including video hypermosaicing to generate from multiple video views of a scene a three-dimensional video mosaic from which diverse virtual video scene images are synthesized, including panoramic, scene interactive and stereoscopic images |
US5926400A (en) | 1996-11-21 | 1999-07-20 | Intel Corporation | Apparatus and method for determining the intensity of a sound in a virtual world |
US6072496A (en) | 1998-06-08 | 2000-06-06 | Microsoft Corporation | Method and system for capturing and representing 3D geometry, color and shading of facial expressions and other animated objects |
US6226003B1 (en) | 1998-08-11 | 2001-05-01 | Silicon Graphics, Inc. | Method for rendering silhouette and true edges of 3-D line drawings with occlusion |
US20020186216A1 (en) * | 2001-06-11 | 2002-12-12 | Baumberg Adam Michael | 3D computer modelling apparatus |
US6496601B1 (en) | 1997-06-23 | 2002-12-17 | Viewpoint Corp. | System and method for asynchronous, adaptive moving picture compression, and decompression |
US20020196256A1 (en) | 2001-05-08 | 2002-12-26 | Hugues Hoppe | Discontinuity edge overdraw |
US6509902B1 (en) | 2000-02-28 | 2003-01-21 | Mitsubishi Electric Research Laboratories, Inc. | Texture filtering for surface elements |
US20030038892A1 (en) | 2001-08-09 | 2003-02-27 | Sidney Wang | Enhancing broadcast of an event with synthetic scene using a depth map |
US6556199B1 (en) | 1999-08-11 | 2003-04-29 | Advanced Research And Technology Institute | Method and apparatus for fast voxelization of volumetric models |
US20030085992A1 (en) | 2000-03-07 | 2003-05-08 | Sarnoff Corporation | Method and apparatus for providing immersive surveillance |
US20030218672A1 (en) | 2002-05-23 | 2003-11-27 | Zhengyou Zhang | Head pose tracking system |
US20040044441A1 (en) | 2002-09-04 | 2004-03-04 | Rakesh Gupta | Environmental reasoning using geometric data structure |
US6781591B2 (en) | 2001-08-15 | 2004-08-24 | Mitsubishi Electric Research Laboratories, Inc. | Blending multiple images using local and global information |
US20040217956A1 (en) * | 2002-02-28 | 2004-11-04 | Paul Besl | Method and system for processing, compressing, streaming, and interactive rendering of 3D color image data |
US20050001832A1 (en) * | 2003-06-23 | 2005-01-06 | Hong Shen | Method for local surface smoothing with application to chest wall nodule segmentation in lung CT data |
US20050013465A1 (en) | 2003-07-10 | 2005-01-20 | Sarnoff Corporation | Method and apparatus for refining target position and size estimates using image and depth data |
US20050017969A1 (en) | 2003-05-27 | 2005-01-27 | Pradeep Sen | Computer graphics rendering using boundary information |
US20050052452A1 (en) * | 2003-09-05 | 2005-03-10 | Canon Europa N.V. | 3D computer surface model generation |
US6968299B1 (en) | 2000-04-14 | 2005-11-22 | International Business Machines Corporation | Method and apparatus for reconstructing a surface using a ball-pivoting algorithm |
US20050280646A1 (en) | 2004-06-18 | 2005-12-22 | Microsoft Corporation | Real-time texture rendering using generalized displacement maps |
US20050285875A1 (en) | 2004-06-28 | 2005-12-29 | Microsoft Corporation | Interactive viewpoint video system and process |
US20060023782A1 (en) | 2004-07-27 | 2006-02-02 | Microsoft Corporation | System and method for off-line multi-view video compression |
US20060028473A1 (en) | 2004-08-03 | 2006-02-09 | Microsoft Corporation | Real-time rendering system and process for interactive viewpoint video |
US20060028489A1 (en) | 2004-08-03 | 2006-02-09 | Microsoft Corporation | Real-time rendering system and process for interactive viewpoint video that was generated using overlapping images of a scene captured from viewpoints forming a grid |
US7023432B2 (en) | 2001-09-24 | 2006-04-04 | Geomagic, Inc. | Methods, apparatus and computer program products that reconstruct surfaces from data point sets |
US20060158509A1 (en) | 2004-10-15 | 2006-07-20 | Kenoyer Michael L | High definition videoconferencing system |
US7096428B2 (en) | 2001-09-28 | 2006-08-22 | Fuji Xerox Co., Ltd. | Systems and methods for providing a spatially indexed panoramic video |
US7106358B2 (en) | 2002-12-30 | 2006-09-12 | Motorola, Inc. | Method, system and apparatus for telepresence communications |
US20060221072A1 (en) | 2005-02-11 | 2006-10-05 | Se Shuen Y S | 3D imaging system |
US20060262856A1 (en) | 2005-05-20 | 2006-11-23 | Microsoft Corporation | Multi-view video coding based on temporal and view decomposition |
US20060290695A1 (en) | 2001-01-05 | 2006-12-28 | Salomie Ioan A | System and method to obtain surface structures of multi-dimensional objects, and to represent those surface structures for animation, transmission and display |
US20070070177A1 (en) | 2005-07-01 | 2007-03-29 | Christensen Dennis G | Visual and aural perspective management for enhanced interactive video telepresence |
US20070237420A1 (en) | 2006-04-10 | 2007-10-11 | Microsoft Corporation | Oblique image stitching |
US20070236656A1 (en) | 2006-04-06 | 2007-10-11 | Jeong Young-Min | Method of modifying color composition for a color-blind person in a mobile displaying apparatus |
US20070263080A1 (en) | 2006-04-20 | 2007-11-15 | Harrell Randy K | System and method for enhancing eye gaze in a telepresence system |
US20080043024A1 (en) | 2006-06-26 | 2008-02-21 | Siemens Corporate Research, Inc. | Method for reconstructing an object subject to a cone beam using a graphic processor unit (gpu) |
US7348976B2 (en) * | 2002-02-06 | 2008-03-25 | Digital Process Ltd. | Three-dimensional shape display program, three-dimensional shape display method, and three-dimensional shape display apparatus |
US20080088626A1 (en) | 2004-12-10 | 2008-04-17 | Kyoto University | Three-Dimensional Image Data Compression System, Method, Program and Recording Medium |
US20080298571A1 (en) | 2007-05-31 | 2008-12-04 | Kurtz Andrew F | Residential video communication system |
US20090033740A1 (en) | 2007-07-31 | 2009-02-05 | Kddi Corporation | Video method for generating free viewpoint video image using divided local regions |
US20090109280A1 (en) | 2007-10-31 | 2009-04-30 | Technion Research And Development Foundation Ltd. | Free viewpoint video |
US20090128548A1 (en) | 2007-11-16 | 2009-05-21 | Sportvision, Inc. | Image repair interface for providing virtual viewpoints |
US7551232B2 (en) | 2005-11-14 | 2009-06-23 | Lsi Corporation | Noise adaptive 3D composite noise reduction |
US20090215533A1 (en) | 2008-02-27 | 2009-08-27 | Gary Zalewski | Methods for capturing depth data of a scene and applying computer actions |
US20090290811A1 (en) | 2008-05-23 | 2009-11-26 | Samsung Electronics Co., Ltd. | System and method for generating a multi-dimensional image |
US20090315978A1 (en) | 2006-06-02 | 2009-12-24 | Eidgenossische Technische Hochschule Zurich | Method and system for generating a 3d representation of a dynamically changing 3d scene |
US20100026712A1 (en) | 2008-07-31 | 2010-02-04 | Stmicroelectronics S.R.L. | Method and system for video rendering, computer program product therefor |
USD610105S1 (en) | 2006-07-10 | 2010-02-16 | Cisco Technology, Inc. | Telepresence system |
US7671893B2 (en) | 2004-07-27 | 2010-03-02 | Microsoft Corp. | System and method for interactive multi-view video |
US20100080448A1 (en) | 2007-04-03 | 2010-04-01 | Wa James Tam | Method and graphical user interface for modifying depth maps |
US7702016B2 (en) | 2004-08-03 | 2010-04-20 | Microsoft Corporation | System and process for compressing and decompressing multiple, layered, video streams of a scene captured from different viewpoints forming a grid using spatial and temporal encoding |
US20100142824A1 (en) | 2007-05-04 | 2010-06-10 | Imec | Method and apparatus for real-time/on-line performing of multi view multimedia applications |
US20100158388A1 (en) | 2008-12-18 | 2010-06-24 | David Bookout | Hardware accelerated silhouette detection |
US20100201681A1 (en) | 2009-02-09 | 2010-08-12 | Microsoft Corporation | Image Editing Consistent with Scene Geometry |
US20100225735A1 (en) | 2009-03-09 | 2010-09-09 | Cisco Technology, Inc. | System and method for providing three dimensional imaging in a network environment |
US20100262628A1 (en) | 2009-04-14 | 2010-10-14 | David William Singer | Method and apparatus for media data transmission |
US20100259595A1 (en) | 2009-04-10 | 2010-10-14 | Nokia Corporation | Methods and Apparatuses for Efficient Streaming of Free View Point Video |
US20100265248A1 (en) | 2009-04-16 | 2010-10-21 | Mccrae James | Multiscale three-dimensional navigation |
US7840638B2 (en) | 2008-06-27 | 2010-11-23 | Microsoft Corporation | Participant positioning in multimedia conferencing |
US20100303303A1 (en) * | 2009-05-29 | 2010-12-02 | Yuping Shen | Methods for recognizing pose and action of articulated objects with collection of planes in motion |
US20100321378A1 (en) | 2009-06-18 | 2010-12-23 | International Business Machines Corporation | Computer Method and Apparatus Providing Interactive Control and Remote Identity Through In-World Proxy |
US20100329358A1 (en) | 2009-06-25 | 2010-12-30 | Microsoft Corporation | Multi-view video compression and streaming |
US20100328475A1 (en) | 2009-06-30 | 2010-12-30 | Cisco Technology, Inc. | Infrared-Aided Depth Estimation |
US20100328437A1 (en) | 2009-06-25 | 2010-12-30 | Siliconfile Technologies Inc. | Distance measuring apparatus having dual stereo camera |
US20110032251A1 (en) | 2009-08-04 | 2011-02-10 | Sai Krishna Pothana | Method and system for texture compression in a system having an avc decoding and a 3d engine |
US20110050859A1 (en) | 2009-09-03 | 2011-03-03 | Technion Research & Development Foundation Ltd. | Devices and methods of generating three dimensional (3d) colored models |
US20110058021A1 (en) | 2009-09-09 | 2011-03-10 | Nokia Corporation | Rendering multiview content in a 3d video system |
US20110084983A1 (en) | 2009-09-29 | 2011-04-14 | Wavelength & Resonance LLC | Systems and Methods for Interaction With a Virtual Environment |
US20110093273A1 (en) | 2009-10-16 | 2011-04-21 | Bowon Lee | System And Method For Determining The Active Talkers In A Video Conference |
US20110096832A1 (en) | 2009-10-23 | 2011-04-28 | Qualcomm Incorporated | Depth map generation techniques for conversion of 2d video data to 3d video data |
US20110122225A1 (en) | 2009-11-23 | 2011-05-26 | General Instrument Corporation | Depth Coding as an Additional Channel to Video Sequence |
US20110169824A1 (en) | 2008-09-29 | 2011-07-14 | Nobutoshi Fujinami | 3d image processing device and method for reducing noise in 3d image processing device |
US20110181685A1 (en) | 2010-01-26 | 2011-07-28 | Polycom, Inc. | Method and Apparatus to Virtualize People with 3D Effect into a Remote Room on a Telepresence Call for True in Person Experience |
US20110211749A1 (en) | 2010-02-28 | 2011-09-01 | Kar Han Tan | System And Method For Processing Video Using Depth Sensor Information |
US8036491B2 (en) | 2005-08-02 | 2011-10-11 | Casio Computer Co., Ltd. | Apparatus and method for aligning images by detecting features |
US20110252320A1 (en) | 2010-04-09 | 2011-10-13 | Nokia Corporation | Method and apparatus for generating a virtual interactive workspace |
US20110261050A1 (en) | 2008-10-02 | 2011-10-27 | Smolic Aljosa | Intermediate View Synthesis and Multi-View Data Signal Extraction |
US20110267344A1 (en) | 2010-04-30 | 2011-11-03 | Liberovision Ag | Method for estimating a pose of an articulated object model |
US20110304619A1 (en) * | 2010-06-10 | 2011-12-15 | Autodesk, Inc. | Primitive quadric surface extraction from unorganized point cloud data |
US20120075303A1 (en) | 2010-09-27 | 2012-03-29 | Johnsson Bjoern | Multi-View Ray Tracing Using Edge Detection and Shader Reuse |
US8156239B1 (en) | 2011-03-09 | 2012-04-10 | Metropcs Wireless, Inc. | Adaptive multimedia renderer |
US20120114039A1 (en) | 2010-11-09 | 2012-05-10 | Sony Computer Entertainment Inc. | Video coding methods and apparatus |
US20120127267A1 (en) | 2010-11-23 | 2012-05-24 | Qualcomm Incorporated | Depth estimation based on global motion |
US20120141016A1 (en) | 2010-12-03 | 2012-06-07 | National University Corporation Nagoya University | Virtual viewpoint image synthesizing method and virtual viewpoint image synthesizing system |
US20120155680A1 (en) | 2010-12-17 | 2012-06-21 | Microsoft Corporation | Virtual audio environment for multidimensional conferencing |
US20130039632A1 (en) | 2011-08-08 | 2013-02-14 | Roy Feinson | Surround video playback |
US8411126B2 (en) | 2010-06-24 | 2013-04-02 | Hewlett-Packard Development Company, L.P. | Methods and systems for close proximity spatial audio rendering |
US8441482B2 (en) * | 2009-09-21 | 2013-05-14 | Caustic Graphics, Inc. | Systems and methods for self-intersection avoidance in ray tracing |
US20130257853A1 (en) * | 2012-04-03 | 2013-10-03 | Ryan Michael SCHMIDT | Volume-preserving smoothing brush |
US20130286204A1 (en) | 2012-04-30 | 2013-10-31 | Convoy Technologies Corp. | Motor vehicle camera and monitoring system |
US20140219550A1 (en) | 2011-05-13 | 2014-08-07 | Liberovision Ag | Silhouette-based pose estimation |
-
2012
- 2012-08-03 US US13/566,796 patent/US9767598B2/en active Active
Patent Citations (97)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5850352A (en) | 1995-03-31 | 1998-12-15 | The Regents Of The University Of California | Immersive video, including video hypermosaicing to generate from multiple video views of a scene a three-dimensional video mosaic from which diverse virtual video scene images are synthesized, including panoramic, scene interactive and stereoscopic images |
US5926400A (en) | 1996-11-21 | 1999-07-20 | Intel Corporation | Apparatus and method for determining the intensity of a sound in a virtual world |
US6496601B1 (en) | 1997-06-23 | 2002-12-17 | Viewpoint Corp. | System and method for asynchronous, adaptive moving picture compression, and decompression |
US6072496A (en) | 1998-06-08 | 2000-06-06 | Microsoft Corporation | Method and system for capturing and representing 3D geometry, color and shading of facial expressions and other animated objects |
US6226003B1 (en) | 1998-08-11 | 2001-05-01 | Silicon Graphics, Inc. | Method for rendering silhouette and true edges of 3-D line drawings with occlusion |
US6556199B1 (en) | 1999-08-11 | 2003-04-29 | Advanced Research And Technology Institute | Method and apparatus for fast voxelization of volumetric models |
US6509902B1 (en) | 2000-02-28 | 2003-01-21 | Mitsubishi Electric Research Laboratories, Inc. | Texture filtering for surface elements |
US20030085992A1 (en) | 2000-03-07 | 2003-05-08 | Sarnoff Corporation | Method and apparatus for providing immersive surveillance |
US6968299B1 (en) | 2000-04-14 | 2005-11-22 | International Business Machines Corporation | Method and apparatus for reconstructing a surface using a ball-pivoting algorithm |
US20060290695A1 (en) | 2001-01-05 | 2006-12-28 | Salomie Ioan A | System and method to obtain surface structures of multi-dimensional objects, and to represent those surface structures for animation, transmission and display |
US20020196256A1 (en) | 2001-05-08 | 2002-12-26 | Hugues Hoppe | Discontinuity edge overdraw |
US20020186216A1 (en) * | 2001-06-11 | 2002-12-12 | Baumberg Adam Michael | 3D computer modelling apparatus |
US20030038892A1 (en) | 2001-08-09 | 2003-02-27 | Sidney Wang | Enhancing broadcast of an event with synthetic scene using a depth map |
US6781591B2 (en) | 2001-08-15 | 2004-08-24 | Mitsubishi Electric Research Laboratories, Inc. | Blending multiple images using local and global information |
US7023432B2 (en) | 2001-09-24 | 2006-04-04 | Geomagic, Inc. | Methods, apparatus and computer program products that reconstruct surfaces from data point sets |
US7096428B2 (en) | 2001-09-28 | 2006-08-22 | Fuji Xerox Co., Ltd. | Systems and methods for providing a spatially indexed panoramic video |
US7348976B2 (en) * | 2002-02-06 | 2008-03-25 | Digital Process Ltd. | Three-dimensional shape display program, three-dimensional shape display method, and three-dimensional shape display apparatus |
US20040217956A1 (en) * | 2002-02-28 | 2004-11-04 | Paul Besl | Method and system for processing, compressing, streaming, and interactive rendering of 3D color image data |
US20030218672A1 (en) | 2002-05-23 | 2003-11-27 | Zhengyou Zhang | Head pose tracking system |
US20040044441A1 (en) | 2002-09-04 | 2004-03-04 | Rakesh Gupta | Environmental reasoning using geometric data structure |
US7106358B2 (en) | 2002-12-30 | 2006-09-12 | Motorola, Inc. | Method, system and apparatus for telepresence communications |
US20050017969A1 (en) | 2003-05-27 | 2005-01-27 | Pradeep Sen | Computer graphics rendering using boundary information |
US20050001832A1 (en) * | 2003-06-23 | 2005-01-06 | Hong Shen | Method for local surface smoothing with application to chest wall nodule segmentation in lung CT data |
US20050013465A1 (en) | 2003-07-10 | 2005-01-20 | Sarnoff Corporation | Method and apparatus for refining target position and size estimates using image and depth data |
US20050052452A1 (en) * | 2003-09-05 | 2005-03-10 | Canon Europa N.V. | 3D computer surface model generation |
US20050280646A1 (en) | 2004-06-18 | 2005-12-22 | Microsoft Corporation | Real-time texture rendering using generalized displacement maps |
US20050285875A1 (en) | 2004-06-28 | 2005-12-29 | Microsoft Corporation | Interactive viewpoint video system and process |
US20050286759A1 (en) | 2004-06-28 | 2005-12-29 | Microsoft Corporation | Interactive viewpoint video system and process employing overlapping images of a scene captured from viewpoints forming a grid |
US7286143B2 (en) | 2004-06-28 | 2007-10-23 | Microsoft Corporation | Interactive viewpoint video employing viewpoints forming an array |
US7671893B2 (en) | 2004-07-27 | 2010-03-02 | Microsoft Corp. | System and method for interactive multi-view video |
US20060023782A1 (en) | 2004-07-27 | 2006-02-02 | Microsoft Corporation | System and method for off-line multi-view video compression |
US20060028489A1 (en) | 2004-08-03 | 2006-02-09 | Microsoft Corporation | Real-time rendering system and process for interactive viewpoint video that was generated using overlapping images of a scene captured from viewpoints forming a grid |
US7702016B2 (en) | 2004-08-03 | 2010-04-20 | Microsoft Corporation | System and process for compressing and decompressing multiple, layered, video streams of a scene captured from different viewpoints forming a grid using spatial and temporal encoding |
US7142209B2 (en) | 2004-08-03 | 2006-11-28 | Microsoft Corporation | Real-time rendering system and process for interactive viewpoint video that was generated using overlapping images of a scene captured from viewpoints forming a grid |
US20060028473A1 (en) | 2004-08-03 | 2006-02-09 | Microsoft Corporation | Real-time rendering system and process for interactive viewpoint video |
US20060158509A1 (en) | 2004-10-15 | 2006-07-20 | Kenoyer Michael L | High definition videoconferencing system |
US20080088626A1 (en) | 2004-12-10 | 2008-04-17 | Kyoto University | Three-Dimensional Image Data Compression System, Method, Program and Recording Medium |
US20060221072A1 (en) | 2005-02-11 | 2006-10-05 | Se Shuen Y S | 3D imaging system |
US20060262856A1 (en) | 2005-05-20 | 2006-11-23 | Microsoft Corporation | Multi-view video coding based on temporal and view decomposition |
US20070070177A1 (en) | 2005-07-01 | 2007-03-29 | Christensen Dennis G | Visual and aural perspective management for enhanced interactive video telepresence |
US8036491B2 (en) | 2005-08-02 | 2011-10-11 | Casio Computer Co., Ltd. | Apparatus and method for aligning images by detecting features |
US7551232B2 (en) | 2005-11-14 | 2009-06-23 | Lsi Corporation | Noise adaptive 3D composite noise reduction |
US20070236656A1 (en) | 2006-04-06 | 2007-10-11 | Jeong Young-Min | Method of modifying color composition for a color-blind person in a mobile displaying apparatus |
US7778491B2 (en) | 2006-04-10 | 2010-08-17 | Microsoft Corporation | Oblique image stitching |
US20070237420A1 (en) | 2006-04-10 | 2007-10-11 | Microsoft Corporation | Oblique image stitching |
US20070263080A1 (en) | 2006-04-20 | 2007-11-15 | Harrell Randy K | System and method for enhancing eye gaze in a telepresence system |
US20090315978A1 (en) | 2006-06-02 | 2009-12-24 | Eidgenossische Technische Hochschule Zurich | Method and system for generating a 3d representation of a dynamically changing 3d scene |
US20080043024A1 (en) | 2006-06-26 | 2008-02-21 | Siemens Corporate Research, Inc. | Method for reconstructing an object subject to a cone beam using a graphic processor unit (gpu) |
USD610105S1 (en) | 2006-07-10 | 2010-02-16 | Cisco Technology, Inc. | Telepresence system |
US20100080448A1 (en) | 2007-04-03 | 2010-04-01 | Wa James Tam | Method and graphical user interface for modifying depth maps |
US20100142824A1 (en) | 2007-05-04 | 2010-06-10 | Imec | Method and apparatus for real-time/on-line performing of multi view multimedia applications |
US20080298571A1 (en) | 2007-05-31 | 2008-12-04 | Kurtz Andrew F | Residential video communication system |
US20090033740A1 (en) | 2007-07-31 | 2009-02-05 | Kddi Corporation | Video method for generating free viewpoint video image using divided local regions |
US20090109280A1 (en) | 2007-10-31 | 2009-04-30 | Technion Research And Development Foundation Ltd. | Free viewpoint video |
US20090128548A1 (en) | 2007-11-16 | 2009-05-21 | Sportvision, Inc. | Image repair interface for providing virtual viewpoints |
US20090215533A1 (en) | 2008-02-27 | 2009-08-27 | Gary Zalewski | Methods for capturing depth data of a scene and applying computer actions |
US20090290811A1 (en) | 2008-05-23 | 2009-11-26 | Samsung Electronics Co., Ltd. | System and method for generating a multi-dimensional image |
US7840638B2 (en) | 2008-06-27 | 2010-11-23 | Microsoft Corporation | Participant positioning in multimedia conferencing |
US8106924B2 (en) | 2008-07-31 | 2012-01-31 | Stmicroelectronics S.R.L. | Method and system for video rendering, computer program product therefor |
US20100026712A1 (en) | 2008-07-31 | 2010-02-04 | Stmicroelectronics S.R.L. | Method and system for video rendering, computer program product therefor |
US20110169824A1 (en) | 2008-09-29 | 2011-07-14 | Nobutoshi Fujinami | 3d image processing device and method for reducing noise in 3d image processing device |
US20110261050A1 (en) | 2008-10-02 | 2011-10-27 | Smolic Aljosa | Intermediate View Synthesis and Multi-View Data Signal Extraction |
US20100158388A1 (en) | 2008-12-18 | 2010-06-24 | David Bookout | Hardware accelerated silhouette detection |
US20100201681A1 (en) | 2009-02-09 | 2010-08-12 | Microsoft Corporation | Image Editing Consistent with Scene Geometry |
US20100225735A1 (en) | 2009-03-09 | 2010-09-09 | Cisco Technology, Inc. | System and method for providing three dimensional imaging in a network environment |
US20100259595A1 (en) | 2009-04-10 | 2010-10-14 | Nokia Corporation | Methods and Apparatuses for Efficient Streaming of Free View Point Video |
US20100262628A1 (en) | 2009-04-14 | 2010-10-14 | David William Singer | Method and apparatus for media data transmission |
US20100265248A1 (en) | 2009-04-16 | 2010-10-21 | Mccrae James | Multiscale three-dimensional navigation |
US20100303303A1 (en) * | 2009-05-29 | 2010-12-02 | Yuping Shen | Methods for recognizing pose and action of articulated objects with collection of planes in motion |
US20100321378A1 (en) | 2009-06-18 | 2010-12-23 | International Business Machines Corporation | Computer Method and Apparatus Providing Interactive Control and Remote Identity Through In-World Proxy |
US20100329358A1 (en) | 2009-06-25 | 2010-12-30 | Microsoft Corporation | Multi-view video compression and streaming |
US20100328437A1 (en) | 2009-06-25 | 2010-12-30 | Siliconfile Technologies Inc. | Distance measuring apparatus having dual stereo camera |
US20100328475A1 (en) | 2009-06-30 | 2010-12-30 | Cisco Technology, Inc. | Infrared-Aided Depth Estimation |
US20110032251A1 (en) | 2009-08-04 | 2011-02-10 | Sai Krishna Pothana | Method and system for texture compression in a system having an avc decoding and a 3d engine |
US20110050859A1 (en) | 2009-09-03 | 2011-03-03 | Technion Research & Development Foundation Ltd. | Devices and methods of generating three dimensional (3d) colored models |
US20110058021A1 (en) | 2009-09-09 | 2011-03-10 | Nokia Corporation | Rendering multiview content in a 3d video system |
US8441482B2 (en) * | 2009-09-21 | 2013-05-14 | Caustic Graphics, Inc. | Systems and methods for self-intersection avoidance in ray tracing |
US20110084983A1 (en) | 2009-09-29 | 2011-04-14 | Wavelength & Resonance LLC | Systems and Methods for Interaction With a Virtual Environment |
US20110093273A1 (en) | 2009-10-16 | 2011-04-21 | Bowon Lee | System And Method For Determining The Active Talkers In A Video Conference |
US20110096832A1 (en) | 2009-10-23 | 2011-04-28 | Qualcomm Incorporated | Depth map generation techniques for conversion of 2d video data to 3d video data |
US20110122225A1 (en) | 2009-11-23 | 2011-05-26 | General Instrument Corporation | Depth Coding as an Additional Channel to Video Sequence |
US20110181685A1 (en) | 2010-01-26 | 2011-07-28 | Polycom, Inc. | Method and Apparatus to Virtualize People with 3D Effect into a Remote Room on a Telepresence Call for True in Person Experience |
US20110211749A1 (en) | 2010-02-28 | 2011-09-01 | Kar Han Tan | System And Method For Processing Video Using Depth Sensor Information |
US20110252320A1 (en) | 2010-04-09 | 2011-10-13 | Nokia Corporation | Method and apparatus for generating a virtual interactive workspace |
US20110267344A1 (en) | 2010-04-30 | 2011-11-03 | Liberovision Ag | Method for estimating a pose of an articulated object model |
US20110304619A1 (en) * | 2010-06-10 | 2011-12-15 | Autodesk, Inc. | Primitive quadric surface extraction from unorganized point cloud data |
US8411126B2 (en) | 2010-06-24 | 2013-04-02 | Hewlett-Packard Development Company, L.P. | Methods and systems for close proximity spatial audio rendering |
US20120075303A1 (en) | 2010-09-27 | 2012-03-29 | Johnsson Bjoern | Multi-View Ray Tracing Using Edge Detection and Shader Reuse |
US20120114039A1 (en) | 2010-11-09 | 2012-05-10 | Sony Computer Entertainment Inc. | Video coding methods and apparatus |
US20120127267A1 (en) | 2010-11-23 | 2012-05-24 | Qualcomm Incorporated | Depth estimation based on global motion |
US20120141016A1 (en) | 2010-12-03 | 2012-06-07 | National University Corporation Nagoya University | Virtual viewpoint image synthesizing method and virtual viewpoint image synthesizing system |
US20120155680A1 (en) | 2010-12-17 | 2012-06-21 | Microsoft Corporation | Virtual audio environment for multidimensional conferencing |
US8156239B1 (en) | 2011-03-09 | 2012-04-10 | Metropcs Wireless, Inc. | Adaptive multimedia renderer |
US20140219550A1 (en) | 2011-05-13 | 2014-08-07 | Liberovision Ag | Silhouette-based pose estimation |
US20130039632A1 (en) | 2011-08-08 | 2013-02-14 | Roy Feinson | Surround video playback |
US20130257853A1 (en) * | 2012-04-03 | 2013-10-03 | Ryan Michael SCHMIDT | Volume-preserving smoothing brush |
US20130286204A1 (en) | 2012-04-30 | 2013-10-31 | Convoy Technologies Corp. | Motor vehicle camera and monitoring system |
Non-Patent Citations (99)
Title |
---|
Abimbola, Kehinde, U.S. Office Action, U.S. Appl. No. 13/588,917, Mar. 4, 2015, pp. 1-36. |
Aliaga, D. G., I. Carlbom, A spatial image hierarchy for compression in image-based-rendering, Proc. of the 2005 Int'l Conf. on Image Processing, ICIP 2005, Sep. 11-14, 2005, pp. 609-612, vol. 1, Genoa, Italy. |
Arsenault, R., C. Ware, Frustum view angle, observer view angle and VE navigation, Proc. of the 5th Symposium on Virtual Reality, Oct. 7-10, 2002, Fortaleza, CE, Brazil. |
Berger, K., K. Ruhl, Y. Schroeder, C. Bruemmer, A. Scholz, M. A. Magnor, Markerless motion capture using multiple color-depth sensors, Proc. of the Vision, Modeling, and Visualization Workshop 2011, VMV 2011, Oct. 4-6, 2011, pp. 317-324, Berlin, Germany. |
Bogomjakov, A., C. Gotsmann, M. Magnor, Free-viewpoint video from depth cameras, Proc. Vision, Modeling and Visualization, Nov. 2006, pp. 89-96. |
Boukerche, A., R. Jarrar, R. W. Pazzi, A novel interactive streaming protocol for image-based 3D virtual environment navigation, Proc. of IEEE Int'l Conf. on Communications, ICC 2009, Jun. 14-18, 2009, pp. 1-6, Dresden, Germany. |
Cain, Leon T., U.S. Final Office Action, U.S. Appl. No. 13/614,852, Apr. 30, 2015, pp. 1-23. |
Cain, Leon T., U.S. Office Action, U.S. Appl. No. 13/614,852, Oct. 31, 2014, pp. 1-19. |
Carranza, J., C. Theobalt, M. A. Magnor, H.-P. Seidel, Free-viewpoint video of human actors, ACM Trans. Graph., Jul. 2003, pp. 569-577, vol. 22, No. 3. |
Chang et al., Principal Component Analysis-based Mesh Decomposition, J. Inf. Sci. Eng., May 2009, vol. 25, No. 4, pp. 971-987. |
Cooke et al., Multiple image view synthesis for free viewpoint video applications, IEEE International Conference on Image Processing, Sep. 2005, vol. 1, pp. 1-4. |
Cooper, O. D., Robust generation of 3D models from video footage of urban scenes, Ph.D Thesis, University of Bristol, Mar. 2005. |
Cooper, Oliver Daniel, "Robust Generation of 3D Models from Video Footage of Urban Scenes", Mar. 2005, pp. 1-219. |
Deering, M., Geometry compression, Proc. of the 22nd Annual Conf. on Comp. Graphics and Interactive Techniques, SIGGRAPH 1995, Aug. 6-11, 1995, pp. 13-20, Los Angeles, CA, USA. |
Do, L., S. Zinger, P.H.N. de With, Quality improving techniques for free-viewpoint DIBR, 3DTV-Conference: The True Vision Capture, Transmission and Display of 3D Video, May 4-6, 2009, pp. 1-4, Potsdam, Germany. |
Eisemann, M., F. Klose, M. A. Magnor, Towards plenoptic Raumzeit reconstruction, Video Processing and Computational Video-International Seminar, Oct. 10-15, 2010, pp. 1-24, Dagstuhl Castle, Germany. |
Eisemann, M., F. Klose, M. A. Magnor, Towards plenoptic Raumzeit reconstruction, Video Processing and Computational Video—International Seminar, Oct. 10-15, 2010, pp. 1-24, Dagstuhl Castle, Germany. |
Eisert, P., Virtual video conferencing using 3D model-assisted image-based rendering, The 2nd IEE European Conf. on Visual Media Production, CVMP 2005, Nov. 30,-Dec. 1, 2005, pp. 185-193. |
Ekmekcioglu, E., B. Gunel, M. Dissanayake, S. T. Worrall, A. M. Kondoz, A scalable multi-view audiovisual entertainment framework with content-aware distribution, 17th IEEE Int'l Conf. on Image Processing, ICIP 2010, Sep. 26-29, 2010, pp. 2401-2404, Hong Kong. |
Fitzgibbon, A. W., Y. Wexler, A. Zisserman, Image-based rendering using image-based priors, 9th IEEE Int'l Conf. on Comp. Vision, ICCV 2003, Oct. 14-17, 2003, pp. 1176-1183, Nice, France. |
Gautier, J., E. Bosc, L. Morin, Representation and coding of 3D video data, Nov. 17, 2010, pp. 1-43. |
Goldlücke, B., Multi-camera reconstruction and rendering for free-viewpoint video, Ph.D. Thesis, Nov. 29, 2006, pp. 1-164, Max-Planck-Institut für Informatik. |
Grau, O., Multi-view 4D reconstruction of human action for entertainment applications, Research and Development White Paper, British Broadcasting Company, Nov. 2011, pp. 1-21. |
Guillemaut, et al., "Joint Multi-Layer Segmentation and Reconstruction for Free-Viewpoint Video Applications", Proceedings of International Journal of Computer Vision, May 2009, pp. 73-100. |
Guillemaut, J.-Y., A. Hilton, Joint multi-layer segmentation and reconstruction for free-viewpoint video applications, Int'l J. of Comp. Vision, May 2011, pp. 73-100, vol. 93, No. 1. |
Hajnik, Daniel F., U.S. Final Office Action, U.S. Appl. No. 13/599,170, Jan. 27, 2015, pp. 1-29. |
Hajnik, Daniel F., U.S. Office Action, U.S. Appl. No. 13/599,170, May 18, 2015, pp. 1-25. |
Hajnik, Daniel F., U.S. Office Action, U.S. Appl. No. 13/599,170, Sep. 30, 2014, pp. 1-25. |
He, Weiming, U.S. Office Action, U.S. Appl. No. 13/790,158, Apr. 28, 2015, pp. 1-27. |
Hornung, A., L. Kobbelt, Interactive pixel-accurate free viewpoint rendering from images with silhouette aware sampling, Comput. Graph. Forum, Dec. 2009, pp. 2090-2103, vol. 28, No. 8. |
Joshi, Sunita, U.S. Office Action, U.S. Appl. No. 13/599,678, Apr. 1, 2015, pp. 1-32. |
Kalvin et al., Superfaces: Polygonal Mesh Simplification with Bounded Error, J. IEEE Comp. Graphics and Applications, May 1996, vol. 16, No. 3, pp. 64-77. |
Kanade et al., Virtualized Reality: Constructing Virtual Worlds from Real Scenes, IEEE Multimedia, Immersive Telepresence, Jan. 1997, vol. 4, No. 1, pp. 34-47. |
K-d-tree.pdf from Wikipedia May 2, 2012 p. 1 introduction of k-d tree. * |
Kilner, J., J. Starck, A. Hilton, A comparative study of free-viewpoint video techniques for sports events, European Conf. on Visual Media Production, Nov. 29-30, 2006, pp. 87-96. |
Kilner, J., J. Starck, J.-Y. Guillemaut, A. Hilton, Objective quality assessment in free-viewpoint video production, Sig. Proc.: Image Comm., Jan. 2009, pp. 3-16, vol. 24, No. 1-2. |
Kim, Y. M., D. Chan, C. Theobalt, S. Thrun, Design and calibration of a multi-view TOF sensor fusion system, IEEE Comp. Society Conf. on Comp. Vision and Pattern Recognition Workshops, CVPRW 2008, Jun. 23-28, 2008, pp. 1-7. |
Kurashima, C. S., R. Yang, A. Lastra, Combining approximate geometry with view-dependent texture mapping-A hybrid approach to 3D video teleconferencing, 15th Brazilian Symposium on Comp. Graphics and Image Processing, SIBGRAPI 2002, Oct. 7-10, 2002, pp. 112-119, Fortaleza-CE, Brazil. |
Kurashima, C. S., R. Yang, A. Lastra, Combining approximate geometry with view-dependent texture mapping—A hybrid approach to 3D video teleconferencing, 15th Brazilian Symposium on Comp. Graphics and Image Processing, SIBGRAPI 2002, Oct. 7-10, 2002, pp. 112-119, Fortaleza-CE, Brazil. |
Kurutepe, E., Civanlar, M.R., Tekalp, A.M., Client-driven selective streaming of multiview video for interactive 3DTV, IEEE Transactions on Circuits and Systems for Video Technology, Nov. 2007, vol. 17, No. 11, pp. 1558-1565. |
Kuster, C., T. Popa, C. Zach, C. Gotsman, M. H. Gross, FreeCam: A hybrid camera system for interactive free-viewpoint video, Proc. of the Vision, Modeling, and Visualization Workshop 2011, VMV 2011, Oct. 4-6, 2011, pp. 17-24, Berlin, Germany. |
Lai, K.-K., Y.-L. Chan, C.-H. Fu, W.-C. Siu, Viewpoint switching in multiview videos using SP-frames, Proc. of the Int'l Conf. on Image Processing, ICIP 2008, Oct. 12-15, 2008, pp. 1776-1779, San Diego, California, USA. |
Lamboray, E., S. Würmlin, M. Waschbüsch, M. H. Gross, H. Pfister, Unconstrained free-viewpoint video coding, Proc. of the 2004 Int'l Conf. on Image Processing, ICIP 2004, Oct. 24-27, 2004, pp. 3261-3264, Singapore. |
Lei, C., Y.-H. Yang, Efficient geometric, photometric, and temporal calibration of an array of unsynchronized video cameras, Sixth Canadian Conf. on Comp. and Robot Vision, CRV 2009, May 25-27, 2009, pp. 162-169, Kelowna, British Columbia, Canada. |
Li, W., Free viewpoint video with image-based rendering, Ph.D Dissertation, May 2010, pp. 1-151, Arizona State University. |
Lipski, C., C. Linz, K. Berger, A. Sellent, M. A. Magnor, Virtual video camera: Image-based viewpoint navigation through space and time, Comput. Graph. Forum, Dec. 2010, pp. 2555-2568, vol. 29, No. 8. |
Liu, et al., "A Point-Cloud-Based Multiview Stereo Algorithm for Free-Viewpoint Video", Proceedings of IEEE Transactions on Visualization and Computer Graphics, May 2010, pp. 407-418. |
Liu, S., K. Kang, J.-P. Tarel, D. B. Cooper, Free-form object reconstruction from silhouettes, occluding edges and texture edges: A unified and robust operator based on duality, IEEE Trans. Pattern Anal. Mach. Intell., Jan. 2008, pp. 131-146, vol. 30, No. 1. |
Liu, Y., Q. Dai, W. Xu, A point-cloud-based multiview stereo algorithm for free-viewpoint video, IEEE Trans. Vis. Comput. Graph., May/Jun. 2010, pp. 407-418, vol. 16, No. 3. |
Lu, Z., Y.-W. Tai, M. Ben-Ezra, M. S. Brown, A framework for ultra high resolution 3D imaging, The Twenty-Third IEEE Conf. on Comp. Vision and Pattern Recognition, CVPR 2010, Jun. 13-18, 2010, pp. 1205-1212, San Francisco, CA, USA. |
Ma, Tze, U.S. Office Action, U.S. Appl. No. 13/598,536, Jun. 26, 2015, pp. 1-16. |
Mamou, K. et al., A simple and efficient approach for 3D mesh approximate convex decomposition, 16th Int'l Conf. on Image Processing, ICIP, Nov. 2009, pp. 3501-3504, Cairo, Egypt. |
Mathis, Normal map workflow, Oct. 18, 2005, http://www.poopinmymouth.com/tutorial/normal-workflow.htm. |
Mathis, Normal map workflow, Oct. 18, 2005, http://www.poopinmymouth.com/tutorial/normal—workflow.htm. |
Matthies, L., M. Okutomi, A Bayesian foundation for active stereo vision, Proc. of SPIE Conf. 1198, Sensor Fusion II: Human and Machine Strategies, Nov. 1989, pp. 62-74. |
Miller, G., High quality novel view rendering from multiple cameras, Doctor of Philosphy, University of Surrey, Centre for Vision, Speech and Signal Processing, School of Electronics and Physical Sciences, Dec. 2007, pp. 1-173. |
Morvan, Y., and C. O'Sullivan, Visual tuning of an image-based rendering algorithm, Proceedings of Eurographics, Oct. 2006, pp. 1-6, Ireland, Dublin. |
Morvan, Y., C. O'Sullivan, Visual Tuning of an Image-Based Rendering Algorithm, Proc. of Eurographics, Oct. 2006, pp. 1-6, Ireland, Dublin. |
Morvan, Y., D. Farin, P. De With, System architecture for free-viewpoint video and 3D-TV, IEEE Transactions on Consumer Electronics, May 2008, pp. 925-932, vol. 54, No. 2. |
Nabeshima, et al., "Frame Rate Stabilization by Variable Resolution Shape Reconstruction for On-line Free-viewpoint Video Generation", Proceedings of Asian Conference on Computer Vision, Jan. 13, 2006, pp. 81-90. |
Nabeshima, R., M. Ueda, D. Arita, R. Taniguchi, Frame rate stabilization by variable resolution shape reconstruction for on-line free-viewpoint video generation, Proc. of the 7th Asian Conf. on Comp. Vision, Jan. 13-16, 2006, pp. 81-90, Hyderabad, India. |
Newcombe et al., Live Dense Reconstruction with a Single Moving Camera, The Twenty-Third IEEE Conf. on Comp. Vision and Pattern Recognition, CVPR 2010, Jun. 2010, pp. 1498-1505, San Francisco, CA, USA. |
Nguyen, Kimbinh, U.S. Final Office Action, U.S. Appl. No. 13/599,436, Jun. 19, 2015, pp. 1-20. |
Nguyen, Kimbinh, U.S. Office Action, U.S. Appl. No. 13/599,436, Mar. 2, 2015, pp. 1-17. |
Nguyen, Phung Hoang Joseph, U.S. Office Action, U.S. Appl. No. 13/602,097, Jun. 1, 2015, pp. 1-9. |
Piatti, D., Time-of-flight cameras: Tests, calibration and multi-frame registration for automatic 3D object reconstruction, 2011, pp. 1-10. |
Pollefeys, et al., "Detailed Real-Time Urban 3D Reconstruction from Video", Proceedings of International Journal of Computer Vision, Jul. 2008, pp. 143-167. |
Pollefeys, M., D. Nistér, J.-M. Frahm, A. Akbarzadeh, P. Mordohai, B. Clipp, C. Engels, D. Gallup, S. J. Kim, P. Merrell, C. Salmi, S. N. Sinha, B. Talton, L. Wang, Q. Yang, H. Stewénius, R. Yang, G. Welch, H. Towles, Detailed real-time urban 3D reconstruction from video, Int'l J. of Comp. Vision, Jul. 2008, pp. 143-167, vol. 78, No. 2-3. |
Rankin, A. L., C. F. Bergh, S. B. Goldberg, P. Bellutta, A. Huertas, L. H. Matthies, Passive perception system for day/night autonomous off-road navigation, Proc. SPIE, Jun. 2, 2005, vol. 5804, Unmanned Ground Vehicle Technology VII, pp. 343-358. |
Rus et al., Analysing the Influence of Vertex Clustering on PCA-Based Dynamic Mesh Compression, Proc. of the 6th Int'l Conf. on Articulated Motion and Deformable Objects, AMDO 2010, Jul. 2010, pp. 55-66, Port d'Andratx, Mallorca, Spain. |
Rusinkiewicz et al., Qsplat: A Multiresolution Point Rendering System for Large Meshes, Proc. of the 27th Annual Conf. on Comp. Graphics, SIGGRAPH 2000, Jul. 23-28, 2000, pp. 343-352, New Orleans, Louisiana, USA. |
Shi, S., W. J. Jeon, K. Nahrstedt, R. H. Campbell, Real-time remote rendering of 3D video for mobile devices, Proc. of the 17th Int'l Conf. on Multimedia 2009, ACM Multimedia 2009, Oct. 19-24, 2009, pp. 391-400, Vancouver, British Columbia, Canada. |
Smolić, A., K. Müller, P. Merkle, T. Rein, M. Kautzner, P. Eisert, T. Wiegand, Free viewpoint video extraction, representation, coding, and rendering, Proc. of the 2004 Int'l Conf. on Image Processing, ICIP 2004, Oct. 24-27, 2004, pp. 3287-3290, vol. 5, Singapore. |
Smolić, A., P. Kauff, Interactive 3D video representation and coding technologies, Invited Paper, Proc. of the IEEE, Special Issue on Advances in Video Coding and Delivery, Jan. 2005, pp. 98-110, vol. 93, No. 1. |
Starck, J., J. Kilner, A. Hilton, A free-viewpoint video renderer, J. Graphics, GPU, & Game Tools, 2009, pp. 57-72, vol. 14, No. 3. |
Starck, J., J. Kilner, A. Hilton, Objective quality assessment in free-viewpoint video production, 3DTV Conf.: The True Vision-Capture, Transmission and Display of 3D Video, May 28-30, 2008, pp. 225-228, Istanbul, Turkey. |
Starck, J., J. Kilner, A. Hilton, Objective quality assessment in free-viewpoint video production, 3DTV Conf.: The True Vision—Capture, Transmission and Display of 3D Video, May 28-30, 2008, pp. 225-228, Istanbul, Turkey. |
Sugden, B., M. Iwanicki, Mega meshes: Modelling, rendering and lighting a world made of 100 billion polygons, Game Developers Conf., Feb. 28,-Mar. 4, 2011, pp. 1-67, San Francisco, CA. |
Teitelbaum, Michael E., U.S. Office Action, U.S. Appl. No. 13/566,877, Jun. 8, 2015, pp. 1-13. |
Theobalt, C., M. Li, M. A. Magnor, H.-P. Seidel, A flexible and versatile studio for synchronized multi-view video recording, Vision, Video, and Graphics, WG 2003, Jul. 10-11, 2003, pp. 9-16, University of Bath, UK. |
Tian, D. P.-L. Lai, P. Lopez, C. Gomila, View synthesis techniques for 3D video, Proc. of the SPIE Applications of Digital Image Processing XXXII, Sep. 2009, pp. 74430T-74430T-11, vol. 7443. |
Tian, et al., "View Synthesis Techniques for 3D Video", Proceedings of SPIE vol. 7443, Sep. 2, 2009, pp. 1-11. |
Vertegaal, R., I. Weevers, C. Sohn, C. Cheung, Gaze-2: Conveying eye contact in group video conferencing using eye-controlled camera direction, Proc. of the 2003 Conf. on Human Factors in Computing Systems, CHI 2003, Apr. 5-10, 2003, pp. 521-528, Ft. Lauderdale, Florida, USA. |
Wei, X., L. Yin, Z. Zhu, Q. Ji, Avatar-mediated face tracking and lip reading for human computer interaction, Proc. of the 12th Acm Int'l Conf. on Multimedia, ACM Multimedia 2004, Oct. 10-16, 2004, pp. 500-503, New York, NY, USA. |
Wikipedia, Hidden surface determination, Apr. 23, 2012, pp. 1-4. |
Wills, Diane M., U.S. Notice of Allowance, U.S. Appl. No. 13/599,263, Aug. 21, 2014, pp. 1-7. |
Wills, Diane M., U.S. Office Action, U.S. Appl. No. 13/599,263, May 29, 2014, pp. 1-19. |
Wu, Chong, U.S. Office Action, U.S. Appl. No. 13/744,885, Feb. 12, 2015, pp. 1-19. |
Würmlin, S., E. Lamboray, M. Gross, 3D video fragments: Dynamic point samples for real-time free-viewpoint video, Computers and Graphics,Feb. 2004, vol. 28, No. 1, pp. 3-14. |
Würmlin, S., E. Lamboray, M. Waschbusch, M. Gross, Dynamic point samples for free-viewpoint video, Proc. of the Picture Coding Symposium, Dec. 15-17, 2004, pp. 6, San Francisco, CA. |
Würmlin, S., E. Lamboray, M. Waschbüsch, P. Kaufman, A. Smolić, M. Gross, Image-space free-viewpoint video, Vision, Modeling, and Visualization, VMV 2005, Nov. 16-18, 2005, pp. 453-460, Erlangen, Germany. |
Yang, Z., Yu, B., Nahrstedt, K., Bajscy, R., A multi-stream adaptation framework for bandwidth management in 3D tele-immersion, May 2006, Proc. of the 2006 Int'l Workshop on Network and operating systems support for digital audio and video, pp. 14. |
Yea, S., A. Vetro, View synthesis prediction for multiview video coding, Sig. Proc.: Image Comm., Jan. 2009, pp. 89-100, vol. 24, No. 1-2. |
Yea, S., A. Vetro, View synthesis prediction for rate-overhead reduction in FTV, 3DTV Conf.: The True Vision-Capture, Transmission and Display of 3D Video, May 28-30, 2008, pp. 145-148, Istanbul, Turkey. |
Yea, S., A. Vetro, View synthesis prediction for rate-overhead reduction in FTV, 3DTV Conf.: The True Vision—Capture, Transmission and Display of 3D Video, May 28-30, 2008, pp. 145-148, Istanbul, Turkey. |
Zhu, Y., A novel view multi-view synthesis approach for free viewpoint video, Int'l Joint Conf. on Artificial Intelligence, JCAI '09, Apr. 25-26, 2009, pp. 88-91, Hainan Island, China. |
Ziegler, G., H. P. A. Lensch, M. Magnor, H.-P. Seidel, Multi-video compression in texture space using 4D SPIHT, 2004 IEEE 6th Workshop on Multimedia Signal Processing, Sep. 29,-Oct. 1, 2004, pp. 39-42, MPI Informatik, Saarbrucken, Germany. |
Zinger, Free-viewpoint depth image based rendering, Preprint submitted to Visual Communication and Image Representation, Jun. 26, 2009, pp. 1-27. |
Zitnick, C. L., S. B. Kang, M. Uyttendaele, S. A. J. Winder, R. Szeliski, High-quality video view interpolation using a layered representation, ACM Trans. Graph., Aug. 2004, pp. 600-608, vol. 23, No. 3. |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160292829A1 (en) * | 2012-06-25 | 2016-10-06 | Yoldas Askan | Method of generating a smooth image from point cloud data |
US10032255B2 (en) * | 2012-06-25 | 2018-07-24 | Yoldas Askan | System for smoothing 3D clouds |
US20190043250A1 (en) * | 2012-06-25 | 2019-02-07 | Yoldas Askan | Method of generating a smooth image from point cloud data |
US20180025496A1 (en) * | 2016-07-19 | 2018-01-25 | Qualcomm Incorporated | Systems and methods for improved surface normal estimation |
US10192345B2 (en) * | 2016-07-19 | 2019-01-29 | Qualcomm Incorporated | Systems and methods for improved surface normal estimation |
US10648832B2 (en) * | 2017-09-27 | 2020-05-12 | Toyota Research Institute, Inc. | System and method for in-vehicle display with integrated object detection |
CN110719497A (en) * | 2018-07-12 | 2020-01-21 | 华为技术有限公司 | Point cloud coding and decoding method and coder-decoder |
CN110719497B (en) * | 2018-07-12 | 2021-06-22 | 华为技术有限公司 | Point cloud coding and decoding method and coder-decoder |
US20220028119A1 (en) * | 2018-12-13 | 2022-01-27 | Samsung Electronics Co., Ltd. | Method, device, and computer-readable recording medium for compressing 3d mesh content |
US11995895B2 (en) * | 2019-06-03 | 2024-05-28 | Nvidia Corporation | Multi-object tracking using correlation filters in video analytics applications |
CN110458780A (en) * | 2019-08-14 | 2019-11-15 | 上海眼控科技股份有限公司 | 3D point cloud data de-noising method, apparatus, computer equipment and readable storage medium storing program for executing |
US11321862B2 (en) | 2020-09-15 | 2022-05-03 | Toyota Research Institute, Inc. | Systems and methods for multi-camera modeling with neural camera networks |
US11494927B2 (en) | 2020-09-15 | 2022-11-08 | Toyota Research Institute, Inc. | Systems and methods for self-supervised depth estimation |
US11508080B2 (en) | 2020-09-15 | 2022-11-22 | Toyota Research Institute, Inc. | Systems and methods for generic visual odometry using learned features via neural camera models |
US11615544B2 (en) | 2020-09-15 | 2023-03-28 | Toyota Research Institute, Inc. | Systems and methods for end-to-end map building from a video sequence using neural camera models |
Also Published As
Publication number | Publication date |
---|---|
US20130321393A1 (en) | 2013-12-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9767598B2 (en) | Smoothing and robust normal estimation for 3D point clouds | |
US9256980B2 (en) | Interpolating oriented disks in 3D space for constructing high fidelity geometric proxies from point clouds | |
CN105701857B (en) | Texturing of 3D modeled objects | |
CN105993034B (en) | Contour completion for enhanced surface reconstruction | |
EP3326156B1 (en) | Consistent tessellation via topology-aware surface tracking | |
CN113902061A (en) | Point cloud completion method and device | |
CN113129352B (en) | Sparse light field reconstruction method and device | |
JP7371691B2 (en) | Point cloud encoding using homography transformation | |
CN113313832B (en) | Semantic generation method and device of three-dimensional model, storage medium and electronic equipment | |
CN111868738B (en) | Cross-device monitoring computer vision system | |
JP2015504559A (en) | Method and apparatus for compression of mirror symmetry based 3D model | |
TW202220440A (en) | Gpcc planar mode and buffer simplification | |
CN115546371A (en) | Point cloud optimization method and system, electronic device and storage medium | |
US8605991B2 (en) | Method for generating visual hulls for 3D objects as sets of convex polyhedra from polygonal silhouettes | |
CN116863078A (en) | Three-dimensional human body model reconstruction method, three-dimensional human body model reconstruction device, electronic equipment and readable medium | |
CN114677439A (en) | Camera pose determination method and device, electronic equipment and storage medium | |
CN114463409A (en) | Method and device for determining image depth information, electronic equipment and medium | |
Heimann et al. | Joint Geometry and Attribute Upsampling of Point Clouds Using Frequency-Selective Models with Overlapped Support | |
CN116012666B (en) | Image generation, model training and information reconstruction methods and devices and electronic equipment | |
CN118314271B (en) | 3D Gaussian rasterization-based rapid high-precision dense reconstruction method and system | |
Park et al. | 3D mesh construction from depth images with occlusion | |
Graber | Realtime 3D reconstruction | |
CN118229552A (en) | Image processing method, electric apparatus, readable storage medium, and program product | |
WO2024217875A1 (en) | Encoding point data indicating a plurality of points in a three-dimensional space | |
WO2024180125A2 (en) | Apparatus and method for rendering multi-path sound diffraction with multi-layer raster maps |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:WINDER, SIMON A. J.;REEL/FRAME:028726/0115 Effective date: 20120802 |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034544/0541 Effective date: 20141014 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |