WO2022123919A1 - Information processing device, information processing method, and information processing program - Google Patents
Information processing device, information processing method, and information processing program Download PDFInfo
- Publication number
- WO2022123919A1 WO2022123919A1 PCT/JP2021/038741 JP2021038741W WO2022123919A1 WO 2022123919 A1 WO2022123919 A1 WO 2022123919A1 JP 2021038741 W JP2021038741 W JP 2021038741W WO 2022123919 A1 WO2022123919 A1 WO 2022123919A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- information processing
- depth map
- map
- difference
- depth
- Prior art date
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 123
- 238000003672 processing method Methods 0.000 title claims abstract description 10
- 238000000605 extraction Methods 0.000 claims abstract description 42
- 239000000284 extract Substances 0.000 claims abstract description 17
- 238000012545 processing Methods 0.000 claims description 32
- 239000003086 colorant Substances 0.000 abstract description 7
- 238000000034 method Methods 0.000 description 29
- 239000000463 material Substances 0.000 description 19
- 230000036544 posture Effects 0.000 description 10
- 238000005516 engineering process Methods 0.000 description 8
- 239000003550 marker Substances 0.000 description 8
- 238000010586 diagram Methods 0.000 description 6
- 235000013305 food Nutrition 0.000 description 6
- 238000004891 communication Methods 0.000 description 4
- 239000002184 metal Substances 0.000 description 4
- 229910052751 metal Inorganic materials 0.000 description 4
- 238000001514 detection method Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 235000013311 vegetables Nutrition 0.000 description 3
- XEEYBQQBJWHFJM-UHFFFAOYSA-N Iron Chemical compound [Fe] XEEYBQQBJWHFJM-UHFFFAOYSA-N 0.000 description 2
- 238000005401 electroluminescence Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 235000013372 meat Nutrition 0.000 description 2
- 150000002739 metals Chemical class 0.000 description 2
- 239000000123 paper Substances 0.000 description 2
- 239000002023 wood Substances 0.000 description 2
- 244000291564 Allium cepa Species 0.000 description 1
- 235000002732 Allium cepa var. cepa Nutrition 0.000 description 1
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 244000000626 Daucus carota Species 0.000 description 1
- 235000002767 Daucus carota Nutrition 0.000 description 1
- 241000196324 Embryophyta Species 0.000 description 1
- 244000088415 Raphanus sativus Species 0.000 description 1
- 235000006140 Raphanus sativus var sativus Nutrition 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000005034 decoration Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000004744 fabric Substances 0.000 description 1
- PCHJSUWPFVWCPO-UHFFFAOYSA-N gold Chemical compound [Au] PCHJSUWPFVWCPO-UHFFFAOYSA-N 0.000 description 1
- 229910052737 gold Inorganic materials 0.000 description 1
- 239000010931 gold Substances 0.000 description 1
- 229910052500 inorganic mineral Inorganic materials 0.000 description 1
- 229910052742 iron Inorganic materials 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000011707 mineral Substances 0.000 description 1
- 230000000737 periodic effect Effects 0.000 description 1
- 229920003002 synthetic resin Polymers 0.000 description 1
- 239000000057 synthetic resin Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/174—Segmentation; Edge detection involving the use of two or more images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/521—Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
Definitions
- This technology relates to information processing devices, information processing methods, and information processing programs.
- RGB camera capable of capturing RGB (Red, Green, Blue) images, it is possible to extract regions having different colors.
- ToF Time Of Flight
- ToF acquires distance information (depth information) by measuring the reflection time of pulsed light applied to an object for each pixel.
- Patent Document 1 For ToF, in order to acquire distance information (depth information) more accurately, the stereo distance calculated according to the stereo method using two images and the accurate distance information of the object are generated using ToF. A technique has been proposed (Patent Document 1).
- This technique has been made in view of these points, and is an information processing device, an information processing method, and an information processing device capable of extracting an area having the same or similar colors in an image of an object or a depth map.
- the purpose is to provide an information processing program.
- the first technique is a difference map generator that generates a difference map from a first depth map for an object acquired by a ToF sensor and a second depth map for the object. It is an information processing apparatus including a region extraction unit that extracts an region on a first depth map based on a difference map.
- the second technique generates a difference map from the first depth map of the object acquired by the ToF sensor and the second depth map of the object, and on the first depth map based on the difference map. It is an information processing method for extracting the area in.
- the third technique generates a difference map from the first depth map for the object acquired by the ToF sensor and the second depth map for the object, and on the first depth map based on the difference map. It is an information processing program that causes a computer to execute an information processing method for extracting an area in.
- the configuration of the information processing system 10 in the first embodiment of the present technology will be described with reference to FIG. 1.
- the information processing system 10 includes an information processing device 100, a ToF sensor 500, and a distance measuring sensor 600.
- the information processing apparatus 100 performs region extraction processing based on a depth map of an object generated by the ToF sensor 500 and an image or depth map of the object generated by the distance measuring sensor 600.
- the object is an object to be subjected to the area extraction process by the information processing apparatus 100, and is two or more objects that can be distinguished from each other, or one object but different materials, materials, and materials on the surface thereof (hereinafter referred to as “objects”). It is an object that has a plurality of areas composed of materials (referred to as materials, etc.).
- Two or more things that can be distinguished are, for example, a "hand (skin) holding a spoon” consisting of a spoon and a hand (skin), and a "food on a plate” consisting of a plate and food.
- a "hand (skin) holding a spoon” consisting of a spoon and a hand (skin)
- a "food on a plate” consisting of a plate and food.
- the surface of the object has multiple areas made of different materials, for example, a tip part (pot) made of metal, a spoon made of a wooden handle, and cardboard.
- the material of the box and its cardboard box is different from that of the box, and the characters, figures, decorations, etc. attached to the surface of the box.
- the information processing apparatus 100 When two or more areas are extracted by the information processing apparatus 100, there are cases where there are two or more objects and each object is extracted as an area, and there is one object but the object. There may be more than one distinguishable area on the surface. Further, there may be two or more objects and two or more distinguishable areas on the surface of the object.
- the ToF sensor 500 is a sensor using ToF that acquires distance information (first depth map) to an object to be processed by the information processing apparatus 100.
- ToF distance information
- iToF indirect Time of Flight
- dToF direct Time of Flight
- the ToF sensor 500 may use either method.
- iToF is a method of obtaining the depth from the phase difference of a periodic signal.
- dToF is a method of measuring the time when a pulsed laser irradiated from a light source is sent and the time when it is returned, and obtaining the depth from the difference.
- ToF a phenomenon called multipath occurs. ToF measures the time it takes for light to be reflected by an object and returned to be received, but depending on the object, the light may not be totally reflected on the surface of the object, but part of it may enter the inside of the object and be reflected inside. After repeating the above steps, the light returns to the ToF sensor 500. Therefore, depending on the object, it takes a long time for the ToF sensor 500 to receive the light, and the distance is detected farther than it actually is. The reflectance of this object varies depending on the material of the object and the like.
- the user needs to generate a first depth map for the object by using the ToF sensor 500 before performing the processing by the information processing apparatus 100.
- the distance measuring sensor 600 is an RGB stereo camera composed of a first RGB camera 610 and a second RGB camera 620.
- the first RGB camera 610 is a camera capable of capturing RGB (Red, Green, Blue) images, and corresponds to the first imaging device within the scope of claims.
- the first RGB image acquired by taking a picture of the first RGB camera 610 corresponds to the first image in the claims.
- the second RGB camera 620 is a camera capable of capturing RGB (Red, Green, Blue) images, and corresponds to the second imaging device within the scope of claims.
- the second RGB image acquired by the shooting of the second RGB camera 620 corresponds to the second image in the claims.
- the distance measuring sensor 600 will be described unless it is necessary to distinguish between the first RGB camera 610 and the second RGB camera 620.
- the user needs to acquire the first RGB image and the second RGB image of the object by using the distance measuring sensor 600 before performing the processing by the information processing apparatus 100.
- the information processing apparatus 100 includes a control unit 150, a storage unit 160, an interface 170, and a display unit 180.
- the control unit 150 is composed of a CPU (Central Processing Unit), a RAM (Random Access Memory), a ROM (Read Only Memory), and the like.
- the CPU executes various processes according to the program stored in the ROM and issues commands to control the entire information processing apparatus 100 and each part thereof.
- the storage unit 160 is a large-capacity storage medium such as a hard disk or a flash memory.
- the storage unit 160 stores programs, calibration data, tables, and the like used in processing in the information processing apparatus 100.
- the interface 170 is an interface for communicating with the ToF sensor 500 and the distance measuring sensor 600.
- the interface 170 may include a wired or wireless communication interface. More specifically, the wired or wireless communication interface is cellular communication such as 3TTE, Wi-Fi, Bluetooth (registered trademark), NFC (Near Field Communication), Ethernet (registered trademark), HDMI (registered trademark). (High-Definition Multimedia Interface), USB (Universal Serial Bus), etc. may be included.
- the interface 170 may include a connection terminal between the devices, a bus in the device, and the like. Also, if the information processing device 100 is distributed across a plurality of devices, the interface 170 may include different types of interfaces for each device. For example, the interface 170 may include both a communication interface and an in-device interface.
- the display unit 180 is a display device composed of, for example, an LCD (Liquid Crystal Display), a PDP (Plasma Display Panel), an organic EL (Electro Luminescence) panel, or the like.
- LCD Liquid Crystal Display
- PDP Plasma Display Panel
- organic EL Electro Luminescence
- the information processing apparatus 100 is configured to include functional blocks such as a depth map generation unit 101, a difference map generation unit 102, an area extraction unit 103, and an image processing unit 104. Each of these units is a function realized by the control unit 150. Further, data and information are transmitted and received between each unit and the ToF sensor 500 and the distance measuring sensor 600 by using the interface 170.
- the depth map generation unit 101 performs triangulation by pattern matching or the like using the first RGB image taken by the first RGB camera 610 constituting the distance measuring sensor 600 and the second RGB image taken by the second RGB camera 620. This will generate a second depth map.
- the difference map generation unit 102 generates a difference map using the first depth map generated by the ToF sensor 500 and the second depth map generated by the depth map generation unit 101.
- the area extraction unit 103 extracts an area on the first depth map using the first depth map and the difference map.
- the image processing unit 104 generates a display image for showing the extraction result by the area extraction unit 103.
- the information processing device 100 is configured as described above.
- the functional block in the information processing apparatus 100 may be realized by executing a program, and the execution of the program may cause a personal computer, a tablet terminal, a smartphone, a server apparatus, or the like to have a function as the information processing apparatus 100.
- the program may be installed in a device such as a personal computer in advance, or may be distributed by download, storage medium, or the like so that the user can install it by himself / herself.
- step S101 the information processing apparatus 100 acquires the first depth map of the object from the ToF sensor 500.
- step S102 the information processing apparatus 100 acquires the first RGB image and the second RGB image from the ranging sensor 600. It should be noted that steps S101 and S102 do not necessarily have to be performed in this order, and may be in the reverse order or may be performed substantially at the same time.
- step S103 the depth map generation unit 101 generates a second depth map from the first RGB image and the second RGB image by performing triangulation by pattern matching or the like using the first calibration data.
- the first calibration data is data showing the relative positional relationship between the first RGB camera 610 and the second RGB camera 620. Since the first RGB camera 610 and the second RGB camera 620 have different viewpoints, it is necessary to match the viewpoints using the first calibration data in order to generate the second depth map.
- the distortion correction data may be included in the first calibration data.
- the first calibration data may be stored in the storage unit 160 in advance, or may be stored in the information processing apparatus 100 in advance. Further, the first calibration data such as an external server may be stored, and the information processing apparatus 100 may access and read the server via the interface 170.
- step S104 the difference map generation unit 102 generates a difference map based on the first depth map and the second depth map.
- the second depth map is projected onto the first depth map using the second calibration data to generate the difference map generation map.
- the second calibration data is data showing the relative positional relationship between the ToF sensor 500 and the distance measuring sensor 600. Since the ToF sensor 500 and the distance measuring sensor 600 are separate bodies and have different viewpoints, the viewpoints are matched with the second calibration data and the second depth map is projected onto the first depth map. Then, the difference map is generated by calculating the difference from the difference map generation map for each pixel constituting the first depth map.
- the second calibration data may be stored in the storage unit 160 in advance, or may be stored in the information processing apparatus 100 in advance. Further, the second calibration data such as an external server may be stored, and the information processing apparatus 100 may access and read the server via the interface 170.
- step S105 the region extraction unit 103 generates a histogram of the difference map and detects a peak from the histogram. Peaks in the histogram are detected, for example, as shown in FIG. 5A. Peaks are detected in each region. For example, if there are two objects, two peaks will be detected. Even if the object is one object, if there are two distinguishable regions on the surface of the object, two peaks are detected.
- step S106 the area extraction unit 103 extracts an area from all the pixels constituting the first depth map by extracting all the pixels having the depth difference of the peak detected from the histogram.
- the region extraction unit 103 extracts a region from the first depth map by extracting all pixels having a depth difference included in a predetermined amount of width centered on the peak detected from the histogram. Extract.
- the width of a predetermined amount centered on the peak is a width set based on the variation in depth due to the performance of the ToF sensor 500 and the distance measuring sensor 600, the variation in depth for each object, and the like, and is a width set in advance based on the information processing apparatus 100. Set in.
- the user may be able to select whether to extract the region based on the peak or the region based on the width of a predetermined amount centered on the peak, or the ToF sensor 500 and the ranging sensor 600 may be selected.
- the information processing apparatus 100 may set the extraction method based on the performance and the like.
- the depth that can be acquired by the ToF sensor 500 deviates from the true value depending on the material of the object and the like due to the above-mentioned multipath. Therefore, by comparing the depth acquired by another method for the same object (distance measuring sensor 600 in this embodiment) and the depth acquired by the ToF sensor 500 and grasping the deviation, the area for each material and the like is obtained. It becomes possible to perform extraction.
- step S107 the image processing unit 104 generates a display image in which the area extracted from the first depth map is drawn.
- the display image is generated by drawing the extracted area on the first depth map.
- the display image may be generated by drawing on any of the second depth map, the first RGB image, and the second RGB image.
- the object is a "hand of a person holding a spoon"
- the spoon is made of wood
- the color is similar to that of a person's hand.
- similar in color means the values of hue, saturation, and lightness, which are the elements that make up the color of the spoon, and the values of hue, saturation, and lightness that make up the color of the hand (skin). Is within a predetermined approximation range, or the RGB value representing the color of the spoon and the RGB value representing the color of the hand (skin) are within the predetermined approximation range.
- the gray image of the object "spoon and hand” is shown in FIG. 6A
- the first depth map of the object is shown in FIG. 6B
- the region extraction unit 103 makes the spoon and the hand different regions. It is assumed that it is extracted as.
- the display image is an image showing that the spoon and the hand extracted as different regions are extracted as different regions by drawing the hands in different colors.
- two or more objects may be extracted, or a plurality of regions composed of different materials or the like on the surface of the object may be extracted although the object is one. It may be done. Further, in some cases, two or more objects are extracted, and two or more distinguishable regions are extracted on the surface of the objects.
- the processing according to the first embodiment is performed. According to the first embodiment, since the region is extracted based on the depth difference, it is possible to extract the region that cannot be extracted only by the RGB image because the colors are the same or similar.
- the information processing apparatus 200 includes functional blocks of a depth map generation unit 101, a difference map generation unit 102, an area extraction unit 103, a type identification unit 201, and an image processing unit 104. It is configured to be prepared. Since the depth map generation unit 101, the difference map generation unit 102, the area extraction unit 103, and the image processing unit 104 are the same as those in the first embodiment, the description thereof will be omitted.
- the type specifying unit 201 associates the depth map acquired by the ToF sensor 500 with the peak of the depth difference in the depth map generated from the image acquired by the distance measuring sensor 600 in advance and the type of the object as shown in FIG.
- the table may be possessed by the type specifying unit 201, or may be stored in the storage unit 160 in advance so that the type specifying unit 201 can read out the table of the storage unit 160. Further, the table may be stored in an external server or the like, and the information processing apparatus 200 may access the server via the interface 170 and read the table.
- the table shown in FIG. 8 is for convenience of explanation, and does not mean that the depth differences of wood, food, skin, and cloth described in FIG. 8 are the values shown in FIG.
- the type is a group of items that have the same properties and morphology according to a certain standard.
- There are various types such as metals, plants, foods, organisms, cloths, synthetic resins, minerals, and papers.
- the types may be further classified, for example, food may be classified into vegetables and meat in more detail, and metals may be further classified into iron, copper, gold and the like.
- steps S101 to S106 are the same as the processes in the first embodiment, the description thereof will be omitted.
- the type identification unit 201 specifies the type of the area extracted by the area extraction unit 103 in step S201.
- the histogram of the object generated by the region extraction unit 103 is shown in FIG. And. Then, the histogram of the table and the object are compared, and when the position of the peak is included in a certain kind of range in the table, it is specified as the kind of the material or the like constituting the region.
- the extracted regions are the region of the tree and the skin.
- the type can be specified as an area.
- step S202 the image processing unit 104 generates a display image in which the region extracted from the first depth map and the type of the region are drawn.
- the display image is generated by drawing the extraction area on the first depth map.
- the display image may be generated by drawing on any of the second depth map, the first RGB image, and the second RGB image.
- the gray image of the object "spoon and hand (skin)" is shown in FIG. 11A
- the first depth map of the object is shown in FIG. 11B
- the spoon and hand (skin) are shown by the region extraction unit 103.
- the display image is an image showing that the spoon and the hand (skin) extracted as different regions are extracted as different regions by drawing them in different colors.
- the type of each specified area is indicated by characters corresponding to each extracted area.
- the user can confirm the extracted area and the type of the area.
- the method of indicating the type of the area in the display image may be the method shown in FIG. 11D, and any method may be used as long as the user can know the type of the area. In FIG. 11D, it is assumed that the spoon is shown in blue and the hand (skin) is shown in red.
- the processing in the second embodiment is performed.
- the type of the material or the like constituting the region can be specified.
- the information processing system 30 includes an information processing device 300 and a ToF sensor 500. Since the ToF sensor 500 is the same as that of the first embodiment, the description thereof will be omitted.
- the object has a region on the surface of the object whose material is different from that of the object and whose color is the same or similar.
- the area may be one that originally exists on the surface of the object, or may be something like a marker consisting of characters or figures attached to the surface of the object by a user or the like. "Similar in color" is similar to that described in the first embodiment.
- the information processing apparatus 300 includes functional blocks of a state estimation unit 301, a depth map generation unit 101, a difference map generation unit 102, an area extraction unit 103, and an image processing unit 104. It is configured to be prepared. Since the depth map generation unit 101, the difference map generation unit 102, the area extraction unit 103, and the image processing unit 104 are the same as those in the first embodiment, the description thereof will be omitted.
- the state estimation unit 301 estimates the state of the object using the three-dimensional shape data.
- the state of the object is the shape, posture and size of the object.
- the shape of the object is one type, the state of the object may be only the posture. Further, the size is not essential, and the state of the object may be a combination of posture and shape.
- the three-dimensional shape data is data showing a plurality of three-dimensional shapes such as a sphere, a cylinder, a cone, a square, a rectangle, a hexagonal column, a triangular pyramid, a triangular prism, and a flat plate as shown in FIGS. 14A to 14I.
- the three-dimensional shape data is not limited to these, and any three-dimensional shape information may be used. In order to enable region detection of more various shapes of objects, it is advisable to prepare more three-dimensional shape data in advance.
- the three-dimensional shape data may be possessed by the state estimation unit 301, or may be stored in the storage unit 160 in advance so that the state estimation unit 301 can read out the three-dimensional shape data of the storage unit 160. Further, the table may be stored in an external server or the like, and the information processing apparatus 300 may access the server via the interface 170 and read the three-dimensional shape data.
- the information processing device 300 is configured as described above.
- step S101 the information processing apparatus 300 acquires the first depth map of the object from the ToF sensor 500, and then in step S301, the state estimation unit 301 estimates the state of the object.
- the posture and size of one 3D shape data at a certain viewpoint in 3D space are tentatively determined, and then the 3D shape viewed from that viewpoint.
- Find the depth map for state estimation about The depth map for state estimation is compared with the first depth map for the object. Then, this comparison process is performed for all possible patterns of postures and a plurality of sizes for each of the three-dimensional shape data, and the one with the smallest depth difference is estimated as the shape, posture, and size of the object.
- the range is predetermined for the size, and the temporary size is determined within the range.
- the depth map will be the same regardless of the posture, but in that case, it is better to select one as a representative.
- step S302 the depth map generation unit 101 adjusts the three-dimensional shape data estimated to be the state of the object in the process in step S301 to the estimated posture of the object according to the estimated size of the object. Based on this, a second depth map of the ToF sensor 500 viewpoint is generated.
- step S303 the difference map generation unit 102 generates a difference map based on the first depth map and the second depth map.
- the difference map is generated by calculating the depth difference from the second depth map at each pixel constituting the first depth map, as in the first embodiment.
- step S304 the area extraction unit 103 generates a histogram of the difference map and detects a peak from the histogram.
- step S305 the area extraction unit 103 extracts an area from all the pixels constituting the first depth map by extracting all the pixels having the depth difference of the peak detected in the histogram.
- the region extraction unit 103 extracts a region by extracting all pixels having a depth difference included in a predetermined amount of width centered on the peak detected in the histogram from the first depth map.
- the method of region extraction is the same as that of the first embodiment.
- step S306 the image processing unit 104 generates a display image in which the extraction area extracted from the first depth map is drawn.
- the display image is generated by drawing the extraction area on the first depth map.
- the display image may be generated by drawing on the second depth map or the three-dimensional shape data.
- the orientation is not determined only by the shape, and a marker is attached to any surface of an object that looks the same when viewed from any direction, up, down, left, or right, and the marker is used in the present technique. By extracting, the direction and surface of the object can be easily grasped.
- a marker character "front” in FIG. 16A
- a marker of a material different from the object is provided as a mark on the front surface of the box which is the object. Then, by extracting and recognizing the marker as an area by the present technique, it is possible to grasp that the surface on which the marker is located is the front surface of the box.
- the marker can be extracted even if the color is the same as or similar to the object, so it does not interfere with the appearance of the object and does not make the viewer of the object aware that the marker is present. You can also do it.
- secret information is described on the surface of the object as an area where the material is different from the object and the color is the same or similar, only by using this technique.
- the secret information can be read.
- secret information can be exchanged without being known by other persons.
- "secret information" is described in characters as an area on the paper that is the object.
- an RGB stereo camera is used as the distance measuring sensor 600, but the distance measuring sensor may be any camera or sensor as long as it can acquire depth information and generate a depth map.
- the distance measuring sensor 600 may be a stereo IR camera composed of two IR (Infrared) cameras, a triangulation by one IR camera and Structured Light, or the like.
- the information processing apparatus 100 generates a depth map by acquiring an image from the ranging sensor 600, but the information processing apparatus 100 generates a depth map in the ranging sensor 600 or an external device so as to acquire it. You may do it.
- the information processing device may operate on the server or the cloud.
- the information processing apparatus receives the first depth map generated by the ToF sensor 500, the image generated by the distance measuring sensor 600, and the like via the network and processes them.
- the present technology can also take the following configurations.
- a difference map generator that generates a difference map from a first depth map of an object acquired by a ToF sensor and a second depth map of the object.
- An information processing device including an area extraction unit that extracts an area on the first depth map based on the difference map.
- the information processing apparatus according to (1) comprising a depth map generation unit that generates the second depth map.
- the depth map generation unit is based on a first image acquired by a first image pickup apparatus constituting a stereo camera as a distance measuring sensor and a second image acquired by a second image pickup apparatus constituting the stereo camera.
- the information processing apparatus according to (2) which generates the second depth map.
- the difference map generation unit generates the difference map by calculating the depth difference between each pixel constituting the first depth map and the second depth map.
- Information processing equipment (5) The information processing apparatus according to any one of (1) to (4), wherein the region extraction unit extracts a region on the first depth map having a depth difference corresponding to a peak in the histogram of the difference map.
- the region extraction unit extracts a region on the first depth map having a depth difference included in a predetermined width centered on a peak in the histogram of the difference map.
- Information processing device is Described in any one of (1) to (3), the difference map generation unit generates the difference map by calculating the depth difference between each pixel constituting the first depth map and the second depth map.
- the information processing apparatus comprising an image processing unit that generates an image indicating an region extracted by the region extraction unit.
- the information processing apparatus according to any one of (1) to (7) comprising a type specifying unit for specifying the type of the object by referring to a table in which the depth difference and the type of the object are previously associated with each other.
- the information processing apparatus further comprising an image processing unit that generates an image indicating the type of the object specified by the area extraction unit and the area extraction unit.
- Information processing device comprising an image processing unit that generates an image indicating an region extracted by the region extraction unit.
- the information processing apparatus (11) The information processing apparatus according to (10), wherein the state of the object is the posture of the object. (12) The information processing apparatus according to (11), wherein the state of the object is the shape of the object. (13) The information processing apparatus according to (11) or (12), wherein the state of the object is the size of the object. (14) The depth map generation unit generates the second depth map for the three-dimensional shape data with respect to the viewpoint of the ToF sensor based on the state of the object. The information processing apparatus according to (10), wherein the difference map generation unit generates the difference map from the first depth map and the second depth map. (15) A difference map is generated from the first depth map of the object acquired by the ToF sensor and the second depth map of the object.
- An information processing method for extracting an area on the first depth map based on the difference map (16) A difference map is generated from the first depth map of the object acquired by the ToF sensor and the second depth map of the object.
- An information processing program that causes a computer to execute an information processing method for extracting an area on the first depth map based on the difference map.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Optics & Photonics (AREA)
- Image Analysis (AREA)
Abstract
Description
<1.第1の実施の形態>
[1-1.情報処理システム10の構成]
[1-2.情報処理装置100の構成]
[1-3.情報処理装置100による処理]
<2.第2の実施の形態>
[2-1.情報処理装置200の構成]
[2-1.情報処理装置200による処理]
<3.第3の実施の形態>
[3-1.情報処理システム30の構成]
[3-2.情報処理装置300の構成]
[3-3.情報処理装置300による処理]
<4.変形例> Hereinafter, embodiments of the present technology will be described with reference to the drawings. The explanation will be given in the following order.
<1. First Embodiment>
[1-1. Configuration of information processing system 10]
[1-2. Configuration of information processing device 100]
[1-3. Processing by information processing device 100]
<2. Second Embodiment>
[2-1. Configuration of information processing device 200]
[2-1. Processing by information processing device 200]
<3. Third Embodiment>
[3-1. Configuration of information processing system 30]
[3-2. Configuration of information processing device 300]
[3-3. Processing by information processing device 300]
<4. Modification example>
[1-1.情報処理システム10の構成]
図1を参照して本技術の第1の実施の形態における情報処理システム10の構成について説明する。情報処理システム10は情報処理装置100、ToFセンサ500、測距センサ600とから構成されている。 <1. First Embodiment>
[1-1. Configuration of information processing system 10]
The configuration of the
次に情報処理装置100の構成について説明する。情報処理装置100は制御部150、記憶部160、インターフェース170、表示部180を備えて構成されている。 [1-2. Configuration of information processing device 100]
Next, the configuration of the
次に図4のフローチャートを参照して情報処理装置100による処理について説明する。 [1-3. Processing by information processing device 100]
Next, processing by the
[2-1.情報処理装置200の構成]
次に本技術の第2の実施の形態について説明する。情報処理システム10の構成は第1の実施の形態と同様である。第2の実施の形態における情報処理装置200は、図7に示すように、デプスマップ生成部101、差分マップ生成部102、領域抽出部103、種類特定部201、画像処理部104という機能ブロックを備えるように構成されている。デプスマップ生成部101、差分マップ生成部102、領域抽出部103、画像処理部104は第1の実施の形態におけるものと同様であるため、その説明を省略する。 <2. Second Embodiment>
[2-1. Configuration of information processing device 200]
Next, a second embodiment of the present technique will be described. The configuration of the
次に図9のフローチャートを参照して第2の実施の形態における情報処理装置200の処理について説明する。 [2-1. Processing by information processing device 200]
Next, the processing of the
[3-1.情報処理システム30の構成]
次に図12を参照して、本技術の第3の実施の形態における情報処理システム30の構成について説明する。情報処理システム30は、情報処理装置300およびToFセンサ500とから構成されている。ToFセンサ500は第1の実施の形態と同様のものであるためその説明を省略する。 <3. Third Embodiment>
[3-1. Configuration of information processing system 30]
Next, with reference to FIG. 12, the configuration of the
次に第3の実施の形態における情報処理装置300の構成について説明する。機能ブロック以外の構成は第1の実施の形態における構成と同様であるため、その説明を省略する。 [3-2. Configuration of information processing device 300]
Next, the configuration of the
次に図15のフローチャートを参照して第3の実施の形態における情報処理装置300の処理について説明する。なお、第1の実施の形態と同様の処理については詳細な説明は省略する。 [3-3. Processing by information processing device 300]
Next, the processing of the
以上、本技術の実施の形態について具体的に説明したが、本技術は上述の実施の形態に限定されるものではなく、本技術の技術的思想に基づく各種の変形が可能である。 <4. Modification example>
Although the embodiment of the present technology has been specifically described above, the present technology is not limited to the above-described embodiment, and various modifications based on the technical idea of the present technology are possible.
(1)
ToFセンサにより取得された対象物についての第1デプスマップと、前記対象物についての第2デプスマップとから差分マップを生成する差分マップ生成部と、
前記差分マップに基づいて前記第1デプスマップ上における領域を抽出する領域抽出部とを備える
情報処理装置。
(2)
前記第2デプスマップを生成するデプスマップ生成部を備える(1)に記載の情報処理装置。
(3)
前記デプスマップ生成部は、測距センサとしてのステレオカメラを構成する第1撮像装置により取得された第1画像と、前記ステレオカメラを構成する第2撮像装置により取得された第2画像とに基づいて前記第2デプスマップを生成する(2)に記載の情報処理装置。
(4)
前記差分マップ生成部は、前記第1デプスマップを構成する各ピクセルにおける前記第2デプスマップとのデプス差を算出することにより前記差分マップを生成する(1)から(3)のいずれかに記載の情報処理装置。
(5)
前記領域抽出部は、前記差分マップのヒストグラムにおけるピークに対応するデプス差を有する前記第1デプスマップ上の領域を抽出する(1)から(4)のいずれかに記載の情報処理装置。
(6)
前記領域抽出部は、前記差分マップのヒストグラムにおけるピークを中心とした所定の幅に含まれるデプス差を有する前記第1デプスマップ上の領域を抽出する(1)から(5)のいずれかに記載の情報処理装置。
(7)
前記領域抽出部が抽出した領域を示す画像を生成する画像処理部を備える(1)から(6)のいずれかに記載の情報処理装置。
(8)
デプス差と前記対象物の種類を予め対応付けたテーブルを参照することにより前記対象物の種類を特定する種類特定部を備える(1)から(7)のいずれかに記載の情報処理装置。
(9)
前記領域抽出部が抽出した領域と前記種類特定部が特定した前記対象物の種類を示す画像を生成する画像処理部を備える(8)に記載の情報処理装置。
(10)
前記第1デプスマップと、3次元形状データから生成される状態推定用デプスマップとに基づいて前記対象物の状態を推定する状態推定部を備える(1)から(9)のいずれかに記載の情報処理装置。
(11)
前記対象物の状態は、前記対象物の姿勢である(10)に記載の情報処理装置。
(12)
前記対象物の状態は、前記対象物の形状である(11)に記載の情報処理装置。
(13)
前記対象物の状態は、前記対象物の大きさである(11)または(12)に記載の情報処理装置。
(14)
前記デプスマップ生成部は、前記対象物の状態に基づいて前記ToFセンサの視点を基準とした前記3次元形状データについての前記第2デプスマップを生成し、
前記差分マップ生成部は、前記第1デプスマップと、前記第2デプスマップとから前記差分マップを生成する(10)に記載の情報処理装置。
(15)
ToFセンサにより取得された対象物についての第1デプスマップと、前記対象物についての第2デプスマップとから差分マップを生成し、
前記差分マップに基づいて前記第1デプスマップ上における領域を抽出する
情報処理方法。
(16)
ToFセンサにより取得された対象物についての第1デプスマップと、前記対象物についての第2デプスマップとから差分マップを生成し、
前記差分マップに基づいて前記第1デプスマップ上における領域を抽出する
情報処理方法をコンピュータに実行させる情報処理プログラム。 The present technology can also take the following configurations.
(1)
A difference map generator that generates a difference map from a first depth map of an object acquired by a ToF sensor and a second depth map of the object.
An information processing device including an area extraction unit that extracts an area on the first depth map based on the difference map.
(2)
The information processing apparatus according to (1), comprising a depth map generation unit that generates the second depth map.
(3)
The depth map generation unit is based on a first image acquired by a first image pickup apparatus constituting a stereo camera as a distance measuring sensor and a second image acquired by a second image pickup apparatus constituting the stereo camera. The information processing apparatus according to (2), which generates the second depth map.
(4)
Described in any one of (1) to (3), the difference map generation unit generates the difference map by calculating the depth difference between each pixel constituting the first depth map and the second depth map. Information processing equipment.
(5)
The information processing apparatus according to any one of (1) to (4), wherein the region extraction unit extracts a region on the first depth map having a depth difference corresponding to a peak in the histogram of the difference map.
(6)
Described in any one of (1) to (5), the region extraction unit extracts a region on the first depth map having a depth difference included in a predetermined width centered on a peak in the histogram of the difference map. Information processing device.
(7)
The information processing apparatus according to any one of (1) to (6), comprising an image processing unit that generates an image indicating an region extracted by the region extraction unit.
(8)
The information processing apparatus according to any one of (1) to (7), comprising a type specifying unit for specifying the type of the object by referring to a table in which the depth difference and the type of the object are previously associated with each other.
(9)
The information processing apparatus according to (8), further comprising an image processing unit that generates an image indicating the type of the object specified by the area extraction unit and the area extraction unit.
(10)
The description according to any one of (1) to (9), comprising a state estimation unit that estimates the state of the object based on the first depth map and the state estimation depth map generated from the three-dimensional shape data. Information processing device.
(11)
The information processing apparatus according to (10), wherein the state of the object is the posture of the object.
(12)
The information processing apparatus according to (11), wherein the state of the object is the shape of the object.
(13)
The information processing apparatus according to (11) or (12), wherein the state of the object is the size of the object.
(14)
The depth map generation unit generates the second depth map for the three-dimensional shape data with respect to the viewpoint of the ToF sensor based on the state of the object.
The information processing apparatus according to (10), wherein the difference map generation unit generates the difference map from the first depth map and the second depth map.
(15)
A difference map is generated from the first depth map of the object acquired by the ToF sensor and the second depth map of the object.
An information processing method for extracting an area on the first depth map based on the difference map.
(16)
A difference map is generated from the first depth map of the object acquired by the ToF sensor and the second depth map of the object.
An information processing program that causes a computer to execute an information processing method for extracting an area on the first depth map based on the difference map.
101・・・デプスマップ生成部
102・・・差分マップ生成部
103・・・領域抽出部
104・・・画像処理部
201・・・種類特定部
301・・・状態推定部
500・・・ToFセンサ
600・・・ステレオカメラ 100, 200, 300 ... Information processing device 101 ... Depth
Claims (16)
- ToFセンサにより取得された対象物についての第1デプスマップと、前記対象物についての第2デプスマップとから差分マップを生成する差分マップ生成部と、
前記差分マップに基づいて前記第1デプスマップ上における領域を抽出する領域抽出部と
を備える情報処理装置。 A difference map generator that generates a difference map from a first depth map of an object acquired by a ToF sensor and a second depth map of the object.
An information processing device including an area extraction unit that extracts an area on the first depth map based on the difference map. - 前記第2デプスマップを生成するデプスマップ生成部を備える
請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising a depth map generation unit that generates the second depth map. - 前記デプスマップ生成部は、測距センサとしてのステレオカメラを構成する第1撮像装置により取得された第1画像と、前記ステレオカメラを構成する第2撮像装置により取得された第2画像とに基づいて前記第2デプスマップを生成する
請求項2に記載の情報処理装置。 The depth map generation unit is based on a first image acquired by a first image pickup apparatus constituting a stereo camera as a distance measuring sensor and a second image acquired by a second image pickup apparatus constituting the stereo camera. The information processing apparatus according to claim 2, wherein the second depth map is generated. - 前記差分マップ生成部は、前記第1デプスマップを構成する各ピクセルにおける前記第2デプスマップとのデプス差を算出することにより前記差分マップを生成する
請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the difference map generation unit generates the difference map by calculating the depth difference between each pixel constituting the first depth map and the second depth map. - 前記領域抽出部は、前記差分マップのヒストグラムにおけるピークに対応するデプス差を有する前記第1デプスマップ上の領域を抽出する
請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the region extraction unit extracts a region on the first depth map having a depth difference corresponding to a peak in the histogram of the difference map. - 前記領域抽出部は、前記差分マップのヒストグラムにおけるピークを中心とした所定の幅に含まれるデプス差を有する前記第1デプスマップ上の領域を抽出する
請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the region extraction unit extracts a region on the first depth map having a depth difference included in a predetermined width centered on a peak in the histogram of the difference map. - 前記領域抽出部が抽出した領域を示す画像を生成する画像処理部を備える
請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising an image processing unit that generates an image indicating an region extracted by the region extraction unit. - デプス差と前記対象物の種類を予め対応付けたテーブルを参照することにより前記対象物の種類を特定する種類特定部を備える
請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising a type specifying unit that specifies the type of the object by referring to a table in which the depth difference and the type of the object are previously associated with each other. - 前記領域抽出部が抽出した領域と前記種類特定部が特定した前記対象物の種類を示す画像を生成する画像処理部を備える
請求項8に記載の情報処理装置。 The information processing apparatus according to claim 8, further comprising an image processing unit that generates an image indicating the type of the object specified by the area extraction unit and the area extraction unit. - 前記第1デプスマップと、3次元形状データから生成される状態推定用デプスマップとに基づいて前記対象物の状態を推定する状態推定部を備える
請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, further comprising a state estimation unit that estimates the state of the object based on the first depth map and the state estimation depth map generated from the three-dimensional shape data. - 前記対象物の状態は、前記対象物の姿勢である
請求項10に記載の情報処理装置。 The information processing device according to claim 10, wherein the state of the object is the posture of the object. - 前記対象物の状態は、前記対象物の形状である
請求項11に記載の情報処理装置。 The information processing apparatus according to claim 11, wherein the state of the object is the shape of the object. - 前記対象物の状態は、前記対象物の大きさである
請求項11に記載の情報処理装置。 The information processing apparatus according to claim 11, wherein the state of the object is the size of the object. - 前記デプスマップ生成部は、前記対象物の状態に基づいて前記ToFセンサの視点を基準とした前記3次元形状データについての前記第2デプスマップを生成し、
前記差分マップ生成部は、前記第1デプスマップと、前記第2デプスマップとから前記差分マップを生成する
請求項10に記載の情報処理装置。 The depth map generation unit generates the second depth map for the three-dimensional shape data with respect to the viewpoint of the ToF sensor based on the state of the object.
The information processing device according to claim 10, wherein the difference map generation unit generates the difference map from the first depth map and the second depth map. - ToFセンサにより取得された対象物についての第1デプスマップと、前記対象物についての第2デプスマップとから差分マップを生成し、
前記差分マップに基づいて前記第1デプスマップ上における領域を抽出する
情報処理方法。 A difference map is generated from the first depth map of the object acquired by the ToF sensor and the second depth map of the object.
An information processing method for extracting an area on the first depth map based on the difference map. - ToFセンサにより取得された対象物についての第1デプスマップと、前記対象物についての第2デプスマップとから差分マップを生成し、
前記差分マップに基づいて前記第1デプスマップ上における領域を抽出する
情報処理方法をコンピュータに実行させる情報処理プログラム。 A difference map is generated from the first depth map of the object acquired by the ToF sensor and the second depth map of the object.
An information processing program that causes a computer to execute an information processing method for extracting an area on the first depth map based on the difference map.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2022568086A JPWO2022123919A1 (en) | 2020-12-11 | 2021-10-20 | |
CN202180081594.1A CN116615748A (en) | 2020-12-11 | 2021-10-20 | Information processing device, information processing method, and information processing program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020-206294 | 2020-12-11 | ||
JP2020206294 | 2020-12-11 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022123919A1 true WO2022123919A1 (en) | 2022-06-16 |
Family
ID=81972853
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/038741 WO2022123919A1 (en) | 2020-12-11 | 2021-10-20 | Information processing device, information processing method, and information processing program |
Country Status (3)
Country | Link |
---|---|
JP (1) | JPWO2022123919A1 (en) |
CN (1) | CN116615748A (en) |
WO (1) | WO2022123919A1 (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015133159A1 (en) * | 2014-03-05 | 2015-09-11 | コニカミノルタ株式会社 | Image processing device, image processing method, and image processing program |
WO2017006574A1 (en) * | 2015-07-03 | 2017-01-12 | オリンパス株式会社 | Image processing device, image determination system, and endoscope system |
JP2018072198A (en) * | 2016-10-31 | 2018-05-10 | 富士通株式会社 | Position posture estimation device, position posture estimation method, and position posture estimation program |
-
2021
- 2021-10-20 CN CN202180081594.1A patent/CN116615748A/en active Pending
- 2021-10-20 JP JP2022568086A patent/JPWO2022123919A1/ja active Pending
- 2021-10-20 WO PCT/JP2021/038741 patent/WO2022123919A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015133159A1 (en) * | 2014-03-05 | 2015-09-11 | コニカミノルタ株式会社 | Image processing device, image processing method, and image processing program |
WO2017006574A1 (en) * | 2015-07-03 | 2017-01-12 | オリンパス株式会社 | Image processing device, image determination system, and endoscope system |
JP2018072198A (en) * | 2016-10-31 | 2018-05-10 | 富士通株式会社 | Position posture estimation device, position posture estimation method, and position posture estimation program |
Also Published As
Publication number | Publication date |
---|---|
CN116615748A (en) | 2023-08-18 |
JPWO2022123919A1 (en) | 2022-06-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10701332B2 (en) | Image processing apparatus, image processing method, image processing system, and storage medium | |
CN105279372B (en) | A kind of method and apparatus of determining depth of building | |
EP2700040B1 (en) | Color channels and optical markers | |
US9396542B2 (en) | Method of estimating imaging device parameters | |
US9639991B2 (en) | Augmented reality provision system | |
US9972091B2 (en) | System and method for detecting object from depth image | |
CN103252778A (en) | Apparatus for estimating the robot pose and method thereof | |
US20190073796A1 (en) | Method and Image Processing System for Determining Parameters of a Camera | |
EP3161725B1 (en) | Color identification using infrared imaging | |
US10984610B2 (en) | Method for influencing virtual objects of augmented reality | |
CN105869115B (en) | A depth image super-resolution method based on kinect2.0 | |
WO2010067486A1 (en) | Image processing device and image processing method | |
WO2020130070A1 (en) | Detecting device, information processing device, detecting method, and information processing program | |
US20200379485A1 (en) | Method for positioning a movable platform, and related device and system | |
CN110942092B (en) | Graphic image recognition method and recognition system | |
CN101674418B (en) | Method for detecting depth of emcee in virtual studio system | |
WO2022123919A1 (en) | Information processing device, information processing method, and information processing program | |
KR101781515B1 (en) | Camera calibration system and method | |
CN105872516A (en) | Method and device for obtaining parallax parameters of three-dimensional film source | |
KR20160024419A (en) | System and Method for identifying stereo-scopic camera in Depth-Image-Based Rendering | |
JP6207968B2 (en) | Forest phase analysis apparatus, forest phase analysis method and program | |
US11657586B1 (en) | Systems and methods for augmented reality viewing based on directly mapped point cloud overlays | |
WO2019198446A1 (en) | Detection device, detection method, information processing device, and information processing program | |
TWM549388U (en) | Object size comparison shopping system | |
CN116883998B (en) | Article labeling method and device based on millimeter wave image and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21903023 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022568086 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202180081594.1 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21903023 Country of ref document: EP Kind code of ref document: A1 |