US20170154433A1 - Obstacle removal using point cloud and depth map data - Google Patents
Obstacle removal using point cloud and depth map data Download PDFInfo
- Publication number
- US20170154433A1 US20170154433A1 US14/956,184 US201514956184A US2017154433A1 US 20170154433 A1 US20170154433 A1 US 20170154433A1 US 201514956184 A US201514956184 A US 201514956184A US 2017154433 A1 US2017154433 A1 US 2017154433A1
- Authority
- US
- United States
- Prior art keywords
- point cloud
- image
- subject matter
- transitory storage
- storage medium
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000004044 response Effects 0.000 claims abstract description 14
- 238000000034 method Methods 0.000 claims description 9
- 230000003287 optical effect Effects 0.000 claims 2
- 239000007787 solid Substances 0.000 claims 2
- 241000283973 Oryctolagus cuniculus Species 0.000 description 12
- 230000006870 function Effects 0.000 description 6
- 241000282320 Panthera leo Species 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/77—Retouching; Inpainting; Scratch removal
-
- G06T7/0079—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G06T7/004—
-
- G06T7/0075—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/557—Depth or shape recovery from multiple images from light fields, e.g. from plenoptic cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/579—Depth or shape recovery from multiple images from motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G06T2207/20144—
Definitions
- the present invention relates generally to obstacle removal using point cloud and depth map data, and more specifically to foreground obstacle removal in digitally captured images using point cloud and depth map data. Even more specifically, the present invention relates to foreground obstacle removal in digitally captured images using point cloud and depth map data to separate desired background subject matter from undesired foreground obstacles.
- composition When a picture (or image) is taken (or captured), its composition may include certain structures or objects that are undesirable. For example, when we take a photo of an outdoor scene, a building, sporting event, wildlife or other animals, or the like there may be structures (i.e., structures composed of geometric features characterized by foreground structure and openings through which background subject matter can be viewed, e.g., a fence or fence-like structure) located between the camera and the scene photographed.
- structures i.e., structures composed of geometric features characterized by foreground structure and openings through which background subject matter can be viewed, e.g., a fence or fence-like structure
- Such structures may be esthetically undesirable to the photographer or artist capturing the image.
- Several embodiments of the invention advantageously address the needs above as well as other needs by providing a system and method for foreground obstacle removal in digitally captured images using point cloud and depth map data to separate desired background subject matter from undesired foreground obstacles.
- the invention can be characterized as a system comprising a non-transitory storage medium, wherein the non-transitory storage medium contains a first image captured at a first position relative to background subject matter, wherein undesirable structure is interposed between the first position and the background subject matter, and a second image captured at a second position, different from the first position, relative to the background subject matter wherein the undesirable structure is interposed between the second position and the background subject matter; a point cloud generation module coupled to the non-transitory storage medium, wherein the point cloud generation module generates a point cloud in response to the first image and the second image, and wherein the point cloud generation module stores the point cloud in the non-transitory storage medium; an identification module coupled to the non-transitory storage medium for retrieving the point cloud and identifying within the point cloud the undesirable structure and the background subject matter; and an image generation module coupled to the identification module for generating a third image in response to the background subject matter within the point cloud, wherein the image generation module stores the third image in the non-transi
- the invention can be characterized as a system comprising a camera comprising a non-transitory storage medium, wherein the non-transitory storage medium contains a first image captured by the camera at a first position relative to background subject matter, wherein undesirable structure is interposed between the first position and the background subject matter, and a second image captured by the camera at a second position, different from the first position, relative to the background subject matter wherein the undesirable structure is interposed between the second position and the background subject matter; a point cloud generation module coupled to the non-transitory storage medium, wherein the point cloud generation module generates a point cloud in response to the first image and the second image, and wherein the point cloud generation module stores the point cloud in the non-transitory storage medium; an identification module coupled to the non-transitory storage medium for retrieving the point cloud and identifying within the point cloud the undesirable structure and the background subject matter; and an image generation module coupled to the identification module for generating a third image in response to the background subject matter within the point cloud, wherein the image generation
- the invention may be characterized as a method comprising providing a non-transitory storage medium, wherein the non-transitory storage medium contains a first image captured at a first position relative to background subject matter, wherein undesirable structure is interposed between the first position and the background subject matter, and a second image captured at a second position, different from the first position, relative to the background subject matter wherein the undesirable structure is interposed between the second position and the background subject matter; generating a point cloud in response to the first image and the second image, and storing the point cloud in the non-transitory storage medium; retrieving the point cloud and identifying within the point cloud the undesirable structure and the background subject matter; generating a third image in response to the background subject matter within the point cloud; and storing the third image in the non-transitory storage medium.
- FIG. 1 is an exemplary image showing esthetically undesirable structure, in the form of a fence-like structure, and background subject matter in the form of a lion.
- FIG. 2 is another exemplary image showing the esthetically undesirable structure, in the form of the fence-like structure, and the background subject matter in the form of a young athlete.
- FIG. 3 is an illustration showing the esthetically undesirable structure, in the form of the fence-like structure, and the background subject matter in the form of a rabbit, and further showing a camera.
- FIG. 4 is an illustration showing the esthetically undesirable structure, in the form of the fence-like structure shown, and the background subject matter in the form of the rabbit, wherein the rabbit is shown as a set of data points comprising a point cloud, such as may be generated by the embodiment of FIG. 3 .
- FIG. 5 is an illustration showing the esthetically undesirable structure having been removed from an image and the background subject matter in the form of the rabbit, wherein the rabbit is shown as a set of data points comprising a point cloud, such as may be generated by the embodiment of FIG. 3 .
- FIG. 6 is four exemplary images, three of which contain the undesirable structure captured from three different positions of a camera, and a fourth of which is generated as a function of the three, wherein the undesirable structure are deleted, such as may be generated by the embodiment of FIG. 3 .
- FIG. 7 is a schematic illustration of the relationship between background subject matter, undesirable structure and positions of a camera, such as may be utilized in the embodiment of FIG. 3 .
- FIG. 8 is a flow chart showing a process for generating an image in which undesirable structure are deleted, such as may be utilized in the embodiment of FIG. 3 .
- a geometric features making up undesirable structure in an image are defined by a predefined geographic shape that permits full view of the background subject matter provided that the background subject matter is viewed from a plurality of positions selected as a function of the predefined geographic shape.
- the plurality of positions must include positions (at least two) defined by a linear relationship between the positions that is not parallel to the one or more parallel structures and a distance between positions that is sufficiently far, as a function of the distance between the camera and the background subject matter and the width of the parallel structures, that the full view of the background subject matter is available in a set of images captured at the plurality of positions.
- an exemplary image 100 is shown showing esthetically undesirable structure 102 , in the form of a fence-like structure, and background subject matter 104 in the form of a lion.
- FIG. 2 another exemplary image 200 is shown showing the esthetically undesirable structure 102 , in the form of the fence-like structure, and the background subject matter 202 in the form of a young athlete.
- an illustration 300 is shown showing the esthetically undesirable structure 102 , in the form of the fence-like structure, and the background subject matter 302 in the form of a rabbit, and further showing a camera 304 .
- the esthetically undesirable structure 102 is shown, as interposed between the background subject matter 302 and the camera 304 .
- the camera 304 is shown in a plurality of positions 306 , 308 , 310 , 312 , 314 , 316 , 318 .
- the camera 304 is positioned in a first position 308 and directed to (i.e., aimed at) the subject matter 302 , with the undesirable structure 102 interposed therein between, so as to take a picture of, i.e., capture an image of, the subject matter 302 .
- a first image is then captured.
- the camera 304 is moved to a second position 316 .
- a line (not shown) between the first position 308 and the second position 316 is not parallel to any part of the undesirable structure 102 . In the case illustrated, this means that the line between the first position 308 and the second position 316 is not parallel to vertical components 320 of the fence, and furthermore, is not parallel to the horizontal components 322 of the fence.
- a line between at least two of the positions 306 , 308 , 310 , 312 , 314 , 316 , 318 must meet these criteria. However, it is not required that all of the lines between all of the positions 306 , 308 , 310 , 312 , 314 , 316 , 318 meet these criteria.
- the line between the first position 308 and the second position 316 may be parallel to the horizontal components 322 of the fence, and the line between the second position 316 and the third position 312 may be parallel to the vertical components 320 of the fence.
- the line between the first position 308 and the third position 312 should not be parallel to either the vertical components 320 of the fence, or the horizontal components 322 of the fence. Note that this exemplary geometry is not illustrated.
- the distance between the positions 306 , 308 , 310 , 312 , 314 , 316 , 318 must be sufficient that the entire background subject matter 302 can be composed by combining the images captured. This is described further herein below in reference to FIG. 7 .
- the first image and the second image are then combined electronically to generate a point cloud (i.e., a three-dimensional representation of the undesirable structure 102 and the subject matter 302 ), as described herein below, and the undesirable structure 102 are electronically identified based on their depth within the point cloud and/or a space between the undesirable structure 102 and the subject matter 302 within the point cloud, and the undesirable structure 102 are then eliminated from (deleted from) the point cloud.
- a resultant image i.e., a two-dimensional image
- a point cloud 402 is generated using heretofore known techniques.
- the point cloud 402 is a set of data points in some coordinate system. In a three-dimensional coordinate system, these points are usually defined using rectangular coordinates, and are often intended to represent the external surface of an object, such as, in this case, the background subject matter 302 in the form of the rabbit.
- point cloud 402 data points are also generated for the undesirable structure 102 , such as, in this case, the fence or the fence-like structure.
- an accurate point cloud of both the undesirable structure and the background subject matter can be generated without any “shadowing” of the background subject matter 302 by the undesirable structure 102 .
- the undesirable structure 102 can be electronically identified and eliminated from (i.e., deleted from) the point cloud 402 .
- an illustration 500 is shown showing the esthetically undesirable structure 102 (represented with dotted lines) having been removed from an image and the background subject matter 302 in the form of the rabbit, wherein the rabbit is shown as a set of data points comprising the point cloud 402 , such as may be generated by the embodiment of FIG. 3 .
- a discernable space between the undesirable structure 102 and the background subject matter 302 is identified and employed by components of the present embodiment in order to delete the undesirable structure 102 from the point cloud 402 .
- the point cloud 402 is then electronically projected onto a plane (not shown) in order to obtain a resultant image of the background subject matter 302 , with the undesirable structure 102 having been deleted.
- the background subject matter 302 may be identified based on a particular depth, i.e., distance from the camera positions (shown in FIG. 3 ), with anything closer to the camera positions being defined as esthetically undesirable structure 102 .
- FIG. 6 four exemplary images 600 , 602 , 604 , 608 are shown, three 600 , 602 , 604 of which contain the undesirable structure 102 captured from three different positions of the camera 312 ( FIG. 3 ), and a fourth 608 of which is generated as a function of the three 600 , 602 , 604 , wherein the undesirable structure 102 is deleted, such as may be generated by the embodiment of FIG. 3 .
- a first 600 of the three images shows the background subject matter 302 , with the undesirable structure 102 interposed between the camera 312 and the background subject matter 302 .
- the undesirable structure 102 is in a first position relative to the background subject matter 302 .
- a second image 602 the undesirable structure 102 is in a second position relative to the background subject matter 302 , this “movement” of the undesirable structure 102 is achieved by moving the camera 312 from a first position 308 to a second position 316 in taking the first image and the second image, respectively.
- the third image 604 of the background subject matter 302 in which the undesirable structure 102 is shown in a third position 312 relative to the background subject matter 302 .
- This further “movement” of the undesirable structure 102 is achieved by moving the camera 312 to a third position during the capture of the third image 604 .
- These three images 600 , 602 , 604 are utilized in the generation of the point cloud 402 ( FIGS. 4 AND 5 ), as described hereinabove.
- the positions 308 , 312 , 316 include at least two positions for which a line between the two positions is not parallel to any of the components of the undesirable structure 102 , for example, the horizontal or vertical components of the fence.
- a discernable space between the undesirable structure 102 and the background subject matter 302 is identified, or a distance from the camera 312 in the positions 308 , 312 , 316 is defined in order to separate the background subject matter 302 from the undesirable structure 102 .
- the undesirable structure 102 is then deleted from the point cloud 402 , and the remaining points in the point cloud 402 are projected onto a plane in order to generate a fourth image.
- the fourth image 608 is an image depicting the background subject matter 302 , in this case a rabbit, wherein the undesirable structure 102 , in this case, a fence, has been deleted.
- the distance from the camera 312 can be defined manually, such as by a slider, which may be displayed on the display of the camera 312 , that is manipulated (such as by a touch screen, arrow keys, joystick, a mechanized slider or the like in order to indicate the distance which is defined as further away from the camera than the undesirable structure 103 and closer to the camera than the background subject matter 302 .
- the distance is selected by the user after the first image 600 , second image 602 and third image 604 are captured.
- the display of the camera may show the resultant image in real time as the user defines the distance, such as by manipulating the slider, thereby allowing the user to see the undesirable structure “disappear” as the distance is defined from a plane in front of the undesirable structure, then through the undesirable structure 103 to a plane between the undesirable structure 103 and the background object matter 302 .
- the distance from the camera 312 can be defined automatically by a processor modified with software within the camera 312 .
- density of the point cloud 402 is evaluated as a function of distance from the camera 312 . This elevation will result in a distinct peak density corresponding to the undesirable structure 103 , and another peak density representing the background subject matter 302 . From within a range of distances between the distinct peak and the other peak, the distance is automatically selected.
- FIG. 7 a schematic illustration 700 is shown of the geometric relationship between background subject matter 302 , undesirable structure 102 and positions 308 , 316 of a camera 312 , such as may be utilized in the embodiment of FIG. 3 .
- the background subject matter 302 and the camera 304 have undesirable structure 102 interposed therein between.
- only one of the vertical members 320 of the fence is schematically depicted. While the camera 304 is in a first position 308 , a portion of the background subject matter 302 is obstructed by the vertical member 320 of the fence, as represented by a first pair of lines 702 .
- the vertical member 320 of the fence has a width W, and is a distance L 2 from the camera 304 during the capture of the first image.
- the background subject matter 302 is at a distance L 1 , which is greater than the distance L 2 , from the camera 304 during the capture of the first image.
- the camera 304 is then moved to a second position 316 with a line 704 between the first position 308 and the second position 316 having a direction that is not parallel to the vertical member 320 illustrated, and being at least a distance D from the first position 308 such that the portion 706 of the background subject matter 302 obstructed by the vertical member 320 having width W is now seen by the camera 304 , as illustrated by the second pair of lines 708 .
- the combination of the first image and the second image provides a complete representation of the background subject matter 302 .
- the undesirable structure 102 can be eliminated by eliminating all points within the above-mentioned point cloud 402 that are no more than a distance L 2 from the camera 304 , while positioning the subject matter 302 at the distance L 1 from the camera 304 .
- FIG. 8 a flow chart 800 showing a process for generating an image in which undesirable structure 102 is deleted, such as may be utilized in the embodiment of FIG. 3 .
- a user of the camera 304 initiates operation of a function or application of the camera adapted to perform the operations described herein.
- the user may then, for example, depress 804 a “start” button and simultaneously therewith move the camera along the line that is not parallel to any of the undesirable structure 102 .
- the camera may then be directed 806 through electronic, software, or mechanical means to automatically capture two or more images.
- the camera 304 then generates 808 the above-described point cloud 402 data from the two or more images and determines 810 , 812 the background subject matter 302 by identifying a discernable space between the undesirable structure 102 and the background subject matter 302 within the point cloud 402 , or defining a distance from the camera 304 during image capture beyond which the background subject matter 302 is present, and closer than which the undesirable structure 102 is present.
- these steps 808 , 810 , and/or 812 may be performed at a location remote from the camera 304 , such as on a server coupled to the camera via a communications network.
- the camera 304 then deletes 814 (or otherwise separates) the undesirable structure 102 from the remainder of the point cloud 402 data, and projects 816 the remaining background subject matter 302 data of the point cloud on to a plane, and outputs 818 this projection as a two-dimensional image.
- This output may be in the form of a digital image file that may, for example, be saved within the camera 304 to a non-transitory medium, or can be transmitted via a communications channel to a remote location at which it can be saved to a non-transitory medium.
- the image may be displayed on a display screen on the camera, or located remotely from the camera, or may be printed on to paper by the camera, or at a location remote from the camera, such as by a photo printer.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Studio Devices (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
- 1. Field of the Invention
- The present invention relates generally to obstacle removal using point cloud and depth map data, and more specifically to foreground obstacle removal in digitally captured images using point cloud and depth map data. Even more specifically, the present invention relates to foreground obstacle removal in digitally captured images using point cloud and depth map data to separate desired background subject matter from undesired foreground obstacles.
- 2. Discussion of the Related Art
- When a picture (or image) is taken (or captured), its composition may include certain structures or objects that are undesirable. For example, when we take a photo of an outdoor scene, a building, sporting event, wildlife or other animals, or the like there may be structures (i.e., structures composed of geometric features characterized by foreground structure and openings through which background subject matter can be viewed, e.g., a fence or fence-like structure) located between the camera and the scene photographed.
- Such structures may be esthetically undesirable to the photographer or artist capturing the image.
- Many image editing or inpainting techniques have been recently developed to remove such undesirable structure and fill the holes left behind in a visually plausible way. However, many existing inpainting techniques require human interactions to label the undesirable structure to be removed. Such labels may take different forms, including strokes, structural boundaries, specified colors, etc. This human labeling step is usually labor intensive (especially in processing a large collection of images) and may require experience and skills.
- Several embodiments of the invention advantageously address the needs above as well as other needs by providing a system and method for foreground obstacle removal in digitally captured images using point cloud and depth map data to separate desired background subject matter from undesired foreground obstacles.
- In one embodiment, the invention can be characterized as a system comprising a non-transitory storage medium, wherein the non-transitory storage medium contains a first image captured at a first position relative to background subject matter, wherein undesirable structure is interposed between the first position and the background subject matter, and a second image captured at a second position, different from the first position, relative to the background subject matter wherein the undesirable structure is interposed between the second position and the background subject matter; a point cloud generation module coupled to the non-transitory storage medium, wherein the point cloud generation module generates a point cloud in response to the first image and the second image, and wherein the point cloud generation module stores the point cloud in the non-transitory storage medium; an identification module coupled to the non-transitory storage medium for retrieving the point cloud and identifying within the point cloud the undesirable structure and the background subject matter; and an image generation module coupled to the identification module for generating a third image in response to the background subject matter within the point cloud, wherein the image generation module stores the third image in the non-transitory storage medium.
- In another embodiment, the invention can be characterized as a system comprising a camera comprising a non-transitory storage medium, wherein the non-transitory storage medium contains a first image captured by the camera at a first position relative to background subject matter, wherein undesirable structure is interposed between the first position and the background subject matter, and a second image captured by the camera at a second position, different from the first position, relative to the background subject matter wherein the undesirable structure is interposed between the second position and the background subject matter; a point cloud generation module coupled to the non-transitory storage medium, wherein the point cloud generation module generates a point cloud in response to the first image and the second image, and wherein the point cloud generation module stores the point cloud in the non-transitory storage medium; an identification module coupled to the non-transitory storage medium for retrieving the point cloud and identifying within the point cloud the undesirable structure and the background subject matter; and an image generation module coupled to the identification module for generating a third image in response to the background subject matter within the point cloud, wherein the image generation module stores the third image in the non-transitory storage medium; and a display device, wherein the display device is coupled to the camera, receives the third image from the camera and displays the third image in a human-perceptible form.
- In a further embodiment, the invention may be characterized as a method comprising providing a non-transitory storage medium, wherein the non-transitory storage medium contains a first image captured at a first position relative to background subject matter, wherein undesirable structure is interposed between the first position and the background subject matter, and a second image captured at a second position, different from the first position, relative to the background subject matter wherein the undesirable structure is interposed between the second position and the background subject matter; generating a point cloud in response to the first image and the second image, and storing the point cloud in the non-transitory storage medium; retrieving the point cloud and identifying within the point cloud the undesirable structure and the background subject matter; generating a third image in response to the background subject matter within the point cloud; and storing the third image in the non-transitory storage medium.
- The above and other aspects, features and advantages of several embodiments of the present invention will be more apparent from the following more particular description thereof, presented in conjunction with the following drawings.
-
FIG. 1 is an exemplary image showing esthetically undesirable structure, in the form of a fence-like structure, and background subject matter in the form of a lion. -
FIG. 2 is another exemplary image showing the esthetically undesirable structure, in the form of the fence-like structure, and the background subject matter in the form of a young athlete. -
FIG. 3 is an illustration showing the esthetically undesirable structure, in the form of the fence-like structure, and the background subject matter in the form of a rabbit, and further showing a camera. -
FIG. 4 is an illustration showing the esthetically undesirable structure, in the form of the fence-like structure shown, and the background subject matter in the form of the rabbit, wherein the rabbit is shown as a set of data points comprising a point cloud, such as may be generated by the embodiment ofFIG. 3 . -
FIG. 5 is an illustration showing the esthetically undesirable structure having been removed from an image and the background subject matter in the form of the rabbit, wherein the rabbit is shown as a set of data points comprising a point cloud, such as may be generated by the embodiment ofFIG. 3 . -
FIG. 6 is four exemplary images, three of which contain the undesirable structure captured from three different positions of a camera, and a fourth of which is generated as a function of the three, wherein the undesirable structure are deleted, such as may be generated by the embodiment ofFIG. 3 . -
FIG. 7 is a schematic illustration of the relationship between background subject matter, undesirable structure and positions of a camera, such as may be utilized in the embodiment ofFIG. 3 . -
FIG. 8 is a flow chart showing a process for generating an image in which undesirable structure are deleted, such as may be utilized in the embodiment ofFIG. 3 . - Corresponding reference characters indicate corresponding components throughout the several views of the drawings. Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures may be exaggerated relative to other elements to help to improve understanding of various embodiments of the present invention. Also, common but well-understood elements that are useful or necessary in a commercially feasible embodiment are often not depicted in order to facilitate a less obstructed view of these various embodiments of the present invention.
- The following description is not to be taken in a limiting sense, but is made merely for the purpose of describing the general principles of exemplary embodiments. The scope of the invention should be determined with reference to the claims.
- Reference throughout this specification to “one embodiment,” “an embodiment,” or similar language means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, appearances of the phrases “in one embodiment,” “in an embodiment,” and similar language throughout this specification may, but do not necessarily, all refer to the same embodiment.
- Furthermore, the described features, structures, or characteristics of the invention may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided, such as examples of programming, software modules, user selections, network transactions, database queries, database structures, hardware modules, hardware circuits, hardware chips, etc., to provide a thorough understanding of embodiments of the invention. One skilled in the relevant art will recognize, however, that the invention can be practiced without one or more of the specific details, or with other methods, components, materials, and so forth. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring aspects of the invention.
- In accordance with the present embodiment, a geometric features making up undesirable structure in an image are defined by a predefined geographic shape that permits full view of the background subject matter provided that the background subject matter is viewed from a plurality of positions selected as a function of the predefined geographic shape. For example, in the case of a fence or fence-like structure composed of one or more sets of parallel structures, e.g., vertical posts, and horizontal cross-members, the plurality of positions must include positions (at least two) defined by a linear relationship between the positions that is not parallel to the one or more parallel structures and a distance between positions that is sufficiently far, as a function of the distance between the camera and the background subject matter and the width of the parallel structures, that the full view of the background subject matter is available in a set of images captured at the plurality of positions.
- Referring first to
FIG. 1 , anexemplary image 100 is shown showing estheticallyundesirable structure 102, in the form of a fence-like structure, andbackground subject matter 104 in the form of a lion. - Referring next of
FIG. 2 , anotherexemplary image 200 is shown showing the estheticallyundesirable structure 102, in the form of the fence-like structure, and thebackground subject matter 202 in the form of a young athlete. - Referring next to
FIG. 3 , anillustration 300 is shown showing the estheticallyundesirable structure 102, in the form of the fence-like structure, and thebackground subject matter 302 in the form of a rabbit, and further showing acamera 304. - The esthetically
undesirable structure 102, is shown, as interposed between thebackground subject matter 302 and thecamera 304. - The
camera 304 is shown in a plurality ofpositions - In operation, the
camera 304 is positioned in afirst position 308 and directed to (i.e., aimed at) thesubject matter 302, with theundesirable structure 102 interposed therein between, so as to take a picture of, i.e., capture an image of, thesubject matter 302. A first image is then captured. Next, thecamera 304 is moved to asecond position 316. A line (not shown) between thefirst position 308 and thesecond position 316 is not parallel to any part of theundesirable structure 102. In the case illustrated, this means that the line between thefirst position 308 and thesecond position 316 is not parallel tovertical components 320 of the fence, and furthermore, is not parallel to thehorizontal components 322 of the fence. - In the event more than two images are captured, a line between at least two of the
positions positions first position 308, asecond position 316, and athird position 312, respectively, the line between thefirst position 308 and thesecond position 316 may be parallel to thehorizontal components 322 of the fence, and the line between thesecond position 316 and thethird position 312 may be parallel to thevertical components 320 of the fence. However, the line between thefirst position 308 and thethird position 312 should not be parallel to either thevertical components 320 of the fence, or thehorizontal components 322 of the fence. Note that this exemplary geometry is not illustrated. - In addition, the distance between the
positions background subject matter 302 can be composed by combining the images captured. This is described further herein below in reference toFIG. 7 . - Returning to the example in which only two images are captured, after the
camera 312 is moved to thesecond position 316, a second image is captured, again with the line between thefirst position 308 and thesecond position 316 being as defined above. - The first image and the second image are then combined electronically to generate a point cloud (i.e., a three-dimensional representation of the
undesirable structure 102 and the subject matter 302), as described herein below, and theundesirable structure 102 are electronically identified based on their depth within the point cloud and/or a space between theundesirable structure 102 and thesubject matter 302 within the point cloud, and theundesirable structure 102 are then eliminated from (deleted from) the point cloud. A resultant image (i.e., a two-dimensional image) is then generated, and output as described herein below. - Referring next to
FIG. 4 , anillustration 400 is shown showing the estheticallyundesirable structure 102, in the form of the fence-like structure, and thebackground subject matter 302 in the form of the rabbit, wherein the rabbit is shown as a set of data points comprising apoint cloud 402, such as may be generated by the embodiment ofFIG. 3 . - In operation, utilizing the first image and the second image, as described in reference to
FIG. 3 , above, apoint cloud 402 is generated using heretofore known techniques. Thepoint cloud 402 is a set of data points in some coordinate system. In a three-dimensional coordinate system, these points are usually defined using rectangular coordinates, and are often intended to represent the external surface of an object, such as, in this case, thebackground subject matter 302 in the form of the rabbit. In addition,point cloud 402 data points are also generated for theundesirable structure 102, such as, in this case, the fence or the fence-like structure. By ensuring that the distance between the first position and the second position, as described above, and as further described below in reference toFIG. 7 , is sufficient, an accurate point cloud of both the undesirable structure and the background subject matter can be generated without any “shadowing” of thebackground subject matter 302 by theundesirable structure 102. Furthermore, because, as viewed in thepoint cloud 402, there is a discernable space between thebackground subject matter 302 and the undesirable structure 102 (and thebackground subject matter 302 is at a greater depth relative to the first position and the second position than the undesirable structure 102), theundesirable structure 102 can be electronically identified and eliminated from (i.e., deleted from) thepoint cloud 402. - Referring next to
FIG. 5 , anillustration 500 is shown showing the esthetically undesirable structure 102 (represented with dotted lines) having been removed from an image and thebackground subject matter 302 in the form of the rabbit, wherein the rabbit is shown as a set of data points comprising thepoint cloud 402, such as may be generated by the embodiment ofFIG. 3 . - As noted above, a discernable space between the
undesirable structure 102 and thebackground subject matter 302 is identified and employed by components of the present embodiment in order to delete theundesirable structure 102 from thepoint cloud 402. - The
point cloud 402 is then electronically projected onto a plane (not shown) in order to obtain a resultant image of thebackground subject matter 302, with theundesirable structure 102 having been deleted. Alternatively, thebackground subject matter 302 may be identified based on a particular depth, i.e., distance from the camera positions (shown inFIG. 3 ), with anything closer to the camera positions being defined as estheticallyundesirable structure 102. - Referring next to
FIG. 6 , fourexemplary images undesirable structure 102 captured from three different positions of the camera 312 (FIG. 3 ), and a fourth 608 of which is generated as a function of the three 600, 602, 604, wherein theundesirable structure 102 is deleted, such as may be generated by the embodiment ofFIG. 3 . - A first 600 of the three images shows the
background subject matter 302, with theundesirable structure 102 interposed between thecamera 312 and thebackground subject matter 302. As will be noted, theundesirable structure 102 is in a first position relative to thebackground subject matter 302. In asecond image 602, theundesirable structure 102 is in a second position relative to thebackground subject matter 302, this “movement” of theundesirable structure 102 is achieved by moving thecamera 312 from afirst position 308 to asecond position 316 in taking the first image and the second image, respectively. Also shown is thethird image 604 of thebackground subject matter 302 in which theundesirable structure 102 is shown in athird position 312 relative to thebackground subject matter 302. This further “movement” of theundesirable structure 102 is achieved by moving thecamera 312 to a third position during the capture of thethird image 604. These threeimages FIGS. 4 AND 5 ), as described hereinabove. And, as will be noted, thepositions undesirable structure 102, for example, the horizontal or vertical components of the fence. Utilizing thepoint cloud 402, a discernable space between theundesirable structure 102 and thebackground subject matter 302 is identified, or a distance from thecamera 312 in thepositions background subject matter 302 from theundesirable structure 102. As referenced above, theundesirable structure 102 is then deleted from thepoint cloud 402, and the remaining points in thepoint cloud 402 are projected onto a plane in order to generate a fourth image. As illustrated, thefourth image 608 is an image depicting thebackground subject matter 302, in this case a rabbit, wherein theundesirable structure 102, in this case, a fence, has been deleted. - In accordance with one variation, the distance from the
camera 312 can be defined manually, such as by a slider, which may be displayed on the display of thecamera 312, that is manipulated (such as by a touch screen, arrow keys, joystick, a mechanized slider or the like in order to indicate the distance which is defined as further away from the camera than the undesirable structure 103 and closer to the camera than thebackground subject matter 302. - Preferably the distance is selected by the user after the
first image 600,second image 602 andthird image 604 are captured. The display of the camera may show the resultant image in real time as the user defines the distance, such as by manipulating the slider, thereby allowing the user to see the undesirable structure “disappear” as the distance is defined from a plane in front of the undesirable structure, then through the undesirable structure 103 to a plane between the undesirable structure 103 and thebackground object matter 302. - In accordance with another variation, the distance from the
camera 312 can be defined automatically by a processor modified with software within thecamera 312. In accordance with this other variation density of thepoint cloud 402 is evaluated as a function of distance from thecamera 312. This elevation will result in a distinct peak density corresponding to the undesirable structure 103, and another peak density representing thebackground subject matter 302. From within a range of distances between the distinct peak and the other peak, the distance is automatically selected. - Referring next to
FIG. 7 , aschematic illustration 700 is shown of the geometric relationship between backgroundsubject matter 302,undesirable structure 102 andpositions camera 312, such as may be utilized in the embodiment ofFIG. 3 . - As shown, the
background subject matter 302 and thecamera 304 haveundesirable structure 102 interposed therein between. In order to simplify this explanation, only one of thevertical members 320 of the fence is schematically depicted. While thecamera 304 is in afirst position 308, a portion of thebackground subject matter 302 is obstructed by thevertical member 320 of the fence, as represented by a first pair oflines 702. Thevertical member 320 of the fence has a width W, and is a distance L2 from thecamera 304 during the capture of the first image. Thebackground subject matter 302 is at a distance L1, which is greater than the distance L2, from thecamera 304 during the capture of the first image. Thecamera 304 is then moved to asecond position 316 with a line 704 between thefirst position 308 and thesecond position 316 having a direction that is not parallel to thevertical member 320 illustrated, and being at least a distance D from thefirst position 308 such that theportion 706 of thebackground subject matter 302 obstructed by thevertical member 320 having width W is now seen by thecamera 304, as illustrated by the second pair oflines 708. As such, the combination of the first image and the second image provides a complete representation of thebackground subject matter 302. Theundesirable structure 102 can be eliminated by eliminating all points within the above-mentionedpoint cloud 402 that are no more than a distance L2 from thecamera 304, while positioning thesubject matter 302 at the distance L1 from thecamera 304. - Referring next to
FIG. 8 , aflow chart 800 showing a process for generating an image in whichundesirable structure 102 is deleted, such as may be utilized in the embodiment ofFIG. 3 . - At the
outset 802, in accordance to one embodiment, a user of thecamera 304 initiates operation of a function or application of the camera adapted to perform the operations described herein. The user may then, for example, depress 804 a “start” button and simultaneously therewith move the camera along the line that is not parallel to any of theundesirable structure 102. Following the pressing of the start button, the camera may then be directed 806 through electronic, software, or mechanical means to automatically capture two or more images. In accordance with the present embodiment, thecamera 304 then generates 808 the above-describedpoint cloud 402 data from the two or more images and determines 810, 812 thebackground subject matter 302 by identifying a discernable space between theundesirable structure 102 and thebackground subject matter 302 within thepoint cloud 402, or defining a distance from thecamera 304 during image capture beyond which thebackground subject matter 302 is present, and closer than which theundesirable structure 102 is present. (Alternatively, thesesteps camera 304, such as on a server coupled to the camera via a communications network.) Thecamera 304 then deletes 814 (or otherwise separates) theundesirable structure 102 from the remainder of thepoint cloud 402 data, andprojects 816 the remainingbackground subject matter 302 data of the point cloud on to a plane, and outputs 818 this projection as a two-dimensional image. (Again, thesesteps 814 and/or 816 may be performed remotely at the server.) This output may be in the form of a digital image file that may, for example, be saved within thecamera 304 to a non-transitory medium, or can be transmitted via a communications channel to a remote location at which it can be saved to a non-transitory medium. In addition, or alternatively, the image may be displayed on a display screen on the camera, or located remotely from the camera, or may be printed on to paper by the camera, or at a location remote from the camera, such as by a photo printer. - While the invention herein disclosed has been described by means of specific embodiments, examples and applications thereof, numerous modifications and variations could be made thereto by those skilled in the art without departing from the scope of the invention set forth in the claims.
Claims (14)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/956,184 US9684965B1 (en) | 2015-12-01 | 2015-12-01 | Obstacle removal using point cloud and depth map data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/956,184 US9684965B1 (en) | 2015-12-01 | 2015-12-01 | Obstacle removal using point cloud and depth map data |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170154433A1 true US20170154433A1 (en) | 2017-06-01 |
US9684965B1 US9684965B1 (en) | 2017-06-20 |
Family
ID=58777067
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/956,184 Active US9684965B1 (en) | 2015-12-01 | 2015-12-01 | Obstacle removal using point cloud and depth map data |
Country Status (1)
Country | Link |
---|---|
US (1) | US9684965B1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210080587A1 (en) * | 2019-09-18 | 2021-03-18 | Topcon Corporation | Surveying data processing device, surveying data processing method, and surveying data processing program |
US11074679B2 (en) * | 2017-02-06 | 2021-07-27 | Huawei Technologies Co., Ltd. | Image correction and display method and device |
US11632489B2 (en) * | 2017-01-31 | 2023-04-18 | Tetavi, Ltd. | System and method for rendering free viewpoint video for studio applications |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6624169B2 (en) * | 2017-07-21 | 2019-12-25 | 株式会社デンソー | Imaging device and display device for vehicle |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160317811A1 (en) * | 2014-09-15 | 2016-11-03 | Second Sight Medical Products, Inc. | Methods and Systems for Detecting Obstacles for a Visual Prosthesis |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010013171A1 (en) | 2008-07-28 | 2010-02-04 | Koninklijke Philips Electronics N.V. | Use of inpainting techniques for image correction |
US20130182184A1 (en) | 2012-01-13 | 2013-07-18 | Turgay Senlet | Video background inpainting |
EP2786582A4 (en) | 2012-02-07 | 2016-03-23 | Nokia Technologies Oy | Object removal from an image |
-
2015
- 2015-12-01 US US14/956,184 patent/US9684965B1/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160317811A1 (en) * | 2014-09-15 | 2016-11-03 | Second Sight Medical Products, Inc. | Methods and Systems for Detecting Obstacles for a Visual Prosthesis |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11632489B2 (en) * | 2017-01-31 | 2023-04-18 | Tetavi, Ltd. | System and method for rendering free viewpoint video for studio applications |
US11665308B2 (en) | 2017-01-31 | 2023-05-30 | Tetavi, Ltd. | System and method for rendering free viewpoint video for sport applications |
US11074679B2 (en) * | 2017-02-06 | 2021-07-27 | Huawei Technologies Co., Ltd. | Image correction and display method and device |
US20210080587A1 (en) * | 2019-09-18 | 2021-03-18 | Topcon Corporation | Surveying data processing device, surveying data processing method, and surveying data processing program |
US11619741B2 (en) * | 2019-09-18 | 2023-04-04 | Topcon Corporation | Surveying data processing device, surveying data processing method, and surveying data processing program |
Also Published As
Publication number | Publication date |
---|---|
US9684965B1 (en) | 2017-06-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109064545B (en) | Method and device for data acquisition and model generation of house | |
US9569873B2 (en) | Automated iterative image-masking based on imported depth information | |
US9684965B1 (en) | Obstacle removal using point cloud and depth map data | |
US8963943B2 (en) | Three-dimensional urban modeling apparatus and method | |
KR20170122725A (en) | Modifying scenes of augmented reality using markers with parameters | |
US9361731B2 (en) | Method and apparatus for displaying video on 3D map | |
JP7391317B2 (en) | Information management device, information management system, information management method, and information management program | |
KR101697713B1 (en) | Method and apparatus for generating intelligence panorama VR(virtual reality) contents | |
US20200226824A1 (en) | Systems and methods for 3d scanning of objects by providing real-time visual feedback | |
CN111046725A (en) | Spatial positioning method based on face recognition and point cloud fusion of surveillance video | |
CN108154058A (en) | Graphic code displaying, the band of position determine method and device | |
CN107690673A (en) | Image processing method and device and server | |
Putra et al. | A proposed combination of photogrammetry, Augmented Reality and Virtual Reality Headset for heritage visualisation | |
EP4111677B1 (en) | Multi-source image data synchronization | |
CN112270736A (en) | Augmented reality processing method and device, storage medium and electronic equipment | |
JP6110780B2 (en) | Additional information display system | |
Basgall et al. | Comparison of lidar and stereo photogrammetric point clouds for change detection | |
KR101038940B1 (en) | System and method for managing image information using object extracted from image | |
KR102271853B1 (en) | Electronic apparatus, image processing method, and computer-readable recording medium | |
WO2017177713A1 (en) | Method and device for adjusting three-dimensional model | |
CN109034214B (en) | Method and apparatus for generating a mark | |
CN111064946A (en) | Video fusion method, system, device and storage medium based on indoor scene | |
CN106611440B (en) | Method and device for extracting live-action picture | |
KR102204721B1 (en) | Method and user terminal for providing AR(Augmented Reality) documentary service | |
KR20180069312A (en) | Method for tracking of object using light field video and apparatus thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TAKAKURA, HIROKI;REEL/FRAME:037261/0438 Effective date: 20151207 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |