US12094151B2 - Image processing system, image processing method, learned model generation method, and data set for learning - Google Patents
Image processing system, image processing method, learned model generation method, and data set for learning Download PDFInfo
- Publication number
- US12094151B2 US12094151B2 US17/422,029 US202017422029A US12094151B2 US 12094151 B2 US12094151 B2 US 12094151B2 US 202017422029 A US202017422029 A US 202017422029A US 12094151 B2 US12094151 B2 US 12094151B2
- Authority
- US
- United States
- Prior art keywords
- dump body
- image
- area
- transporting material
- dump
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000012545 processing Methods 0.000 title claims abstract description 26
- 238000003672 processing method Methods 0.000 title claims description 5
- 238000000034 method Methods 0.000 title description 25
- 239000000463 material Substances 0.000 claims abstract description 58
- 230000011218 segmentation Effects 0.000 claims description 47
- 238000010586 diagram Methods 0.000 description 14
- 210000002569 neuron Anatomy 0.000 description 12
- 238000013528 artificial neural network Methods 0.000 description 10
- 239000013256 coordination polymer Substances 0.000 description 8
- 239000000284 extract Substances 0.000 description 8
- 230000008878 coupling Effects 0.000 description 5
- 238000010168 coupling process Methods 0.000 description 5
- 238000005859 coupling reaction Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 238000003709 image segmentation Methods 0.000 description 5
- 230000009471 action Effects 0.000 description 4
- 238000009412 basement excavation Methods 0.000 description 4
- 238000011156 evaluation Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 239000010720 hydraulic oil Substances 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000012806 monitoring device Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000002787 reinforcement Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 230000032258 transport Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/20—Finite element generation, e.g. wire-frame surface description, tesselation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Definitions
- the present invention relates to an image processing system, an image processing method, a learned model generation method, and a data set for learning.
- Patent Document 1 discloses a technique of extracting an edge from an image captured by an image-capturing device provided in a hydraulic excavator to specify a position of a dump body of a dump truck.
- An object of the present invention is to provide an image processing system, an image processing method, a learned model generation method, and a data set for learning capable of improving robustness of processing of specifying a loading/unloading target of a transporting material from a captured image.
- an image processing system includes a data acquisition unit configured to acquire a captured image showing a loading/unloading target of a transporting material of a work machine, an area-specifying unit configured to specify an area including the loading/unloading target from the captured image, and a loading/unloading target-specifying unit configured to specify at least one predetermined surface of the loading/unloading target from the area including the loading/unloading target.
- the image processing system can improve the robustness of the processing of specifying the loading/unloading target.
- FIG. 1 is a diagram showing a configuration of a loading site according to one embodiment.
- FIG. 2 is an external view of a hydraulic excavator according to one embodiment.
- FIG. 3 is a schematic block diagram showing a configuration of a control device according to a first embodiment.
- FIG. 4 is a diagram showing an example of image segmentation according to the first embodiment.
- FIG. 5 is a diagram showing an example of a configuration of a neural network.
- FIG. 6 is an example of guidance information.
- FIG. 7 is a flowchart showing a display method of the guidance information by the control device according to the first embodiment.
- FIG. 8 is a flowchart showing a learning method of a segmentation model according to the first embodiment.
- FIG. 9 is a schematic block diagram showing a configuration of a control device according to a second embodiment.
- FIG. 10 is a flowchart showing a control method of a hydraulic excavator according to the second embodiment.
- FIG. 11 is a schematic block diagram showing a configuration of a control device according to a third embodiment.
- FIG. 12 is an example of guidance information according to the third embodiment.
- FIG. 13 is a flowchart showing a display method of guidance information by the control device according to the third embodiment.
- FIG. 14 is a diagram showing an example of image segmentation according to a fifth embodiment.
- FIG. 15 is a flowchart showing a display method of guidance information by a control device according to the fifth embodiment.
- FIG. 1 is a diagram showing a configuration of a loading site according to one embodiment.
- a hydraulic excavator 100 which is a loading machine and a dump truck 200 which is a carrying vehicle are deployed.
- the hydraulic excavator 100 scoops a transporting material L such as earth from a construction site and loads the transporting material on the dump truck 200 .
- the dump truck 200 transports the transporting material L loaded by the hydraulic excavator 100 to a predetermined dump site.
- the dump truck 200 includes a dump body 210 , which is a container for accommodating the transporting material L.
- the dump body 210 is an example of a loading/unloading target of the transporting material L.
- the loading site is an example of a site. The site is land where the loading machine is operated.
- FIG. 2 is an external view of a hydraulic excavator according to one embodiment.
- the hydraulic excavator 100 includes work equipment 110 that is hydraulically operated, a swing body 120 that supports the work equipment 110 , and an undercarriage 130 that supports the swing body 120 .
- the work equipment 110 includes a boom 111 , an arm 112 , a bucket 113 , a boom cylinder 114 , an arm cylinder 115 , and a bucket cylinder 116 .
- the boom 111 is a column that supports the arm 112 and the bucket 113 .
- a base end portion of the boom 111 is attached to a front portion of the swing body 120 through a boom pin P 1 .
- the arm 112 connects the boom 111 and the bucket 113 .
- a base end portion of the arm 112 is attached to a tip end portion of the boom 111 through an arm pin P 2 .
- the bucket 113 is a container having a blade for excavating earth or the like. A base end portion of the bucket 113 is attached to a tip end portion of the arm 112 through a bucket pin P 3 .
- the boom cylinder 114 is a hydraulic cylinder for operating the boom 111 .
- a base end portion of the boom cylinder 114 is attached to the swing body 120 .
- a tip end portion of the boom cylinder 114 is attached to the boom 111 .
- the arm cylinder 115 is a hydraulic cylinder for driving the arm 112 .
- a base end portion of the arm cylinder 115 is attached to the boom 111 .
- a tip end portion of the arm cylinder 115 is attached to the arm 112 .
- the bucket cylinder 116 is a hydraulic cylinder for driving the bucket 113 .
- a base end portion of the bucket cylinder 116 is attached to the arm 112 .
- a tip end portion of the bucket cylinder 116 is attached to the bucket 113 .
- the swing body 120 is provided with a cab 121 in which an operator rides.
- the cab 121 is provided in front of the swing body 120 and on a left-side (+Y side) of the work equipment 110 .
- the hydraulic excavator 100 includes a work equipment position detector 122 , a position azimuth direction calculator 123 , an inclination detector 124 , a stereo camera 125 , an operation device 126 , a control device 127 , and a display device 128 .
- the work equipment position detector 122 detects an angle of the work equipment 110 .
- the work equipment position detector 122 according to a first embodiment is a stroke detector that detects a stroke length of each of the boom cylinder 114 , the arm cylinder 115 , and the bucket cylinder 116 . Accordingly, the control device 127 described below can detect the angle of the work equipment 110 based on the stroke length of each of the boom cylinder 114 , the arm cylinder 115 , and the bucket cylinder 116 .
- the present invention is not limited thereto in another embodiment.
- An angle detector such as an IMU, a rotary encoder, or a level may be used as the work equipment position detector 122 , instead of the stroke detector or in combination with the stroke detector.
- the position azimuth direction calculator 123 calculates a position of the swing body 120 and an azimuth direction in which the swing body 120 faces.
- the position azimuth direction calculator 123 includes two receivers that receive positioning signals from artificial satellites configuring a global navigation satellite system (GNSS). Each of the two receivers is installed at a different position of the swing body 120 .
- the position azimuth direction calculator 123 detects a position of a representative point (origin of a vehicle body coordinate system) of the swing body 120 in a site coordinate system based on the positioning signal received by at least one receiver.
- the position azimuth direction calculator 123 calculates the azimuth direction of the swing body 120 based on positions indicated by the two detected positioning signals and grounding positions of the two receivers, using each of the positioning signals received by the two receivers.
- the position azimuth direction calculator 123 may include another device such as a rotary encoder or an IMU, instead of the two receivers to measure the azimuth direction of the swing body 120 .
- the inclination detector 124 measures acceleration and angular velocity of the swing body 120 to detect an inclination of the swing body 120 (for example, a roll representing rotation with respect to the X-axis, a pitch representing rotation with respect to the Y-axis, and a yaw representing rotation with respect to the Z-axis) based on the measurement result.
- the inclination detector 124 is installed, for example, on a lower surface of the cab 121 .
- an inertial measurement unit (IMU) which is an inertial measurement device can be used as the inclination detector 124 .
- the stereo camera 125 is provided on an upper portion of the cab 121 .
- the stereo camera 125 is installed in front (+X direction) of and above (+Z direction) the cab 121 .
- the stereo camera 125 captures an image in front (+X direction) of the cab 121 through a windshield on a front surface of the cab 121 .
- the stereo camera 125 includes at least one pair of cameras.
- the operation device 126 is provided inside the cab 121 .
- the operation device 126 is operated by the operator to supply hydraulic oil to an actuator of the work equipment 110 .
- the hydraulic oil is supplied to the boom cylinder 114 , the arm cylinder 115 , the bucket cylinder 116 , and a swing motor (not shown) according to an operation amount of the operation device 126 , and the work equipment 110 and the swing body 120 are driven.
- the control device 127 acquires information from the work equipment position detector 122 , the position azimuth direction calculator 123 , the inclination detector 124 , and the stereo camera 125 to generate guidance information indicating a positional relationship between the bucket 113 and the dump body of the dump truck 200 .
- the display device 128 displays the guidance information generated by the control device 127 .
- the hydraulic excavator 100 may not include the work equipment position detector 122 , the position azimuth direction calculator 123 , the inclination detector 124 , the stereo camera 125 , and the display device 128 .
- the stereo camera 125 includes a right-side camera 1251 and a left-side camera 1252 .
- Cameras using a charge-coupled device (CCD) sensor and a complementary metal-oxide-semiconductor (CMOS) sensor are exemplary examples of each camera.
- the right-side camera 1251 and the left-side camera 1252 are installed at an interval in a left-right direction (Y-axis direction) such that optical axes of the cameras are substantially parallel to a floor surface of the cab 121 .
- the stereo camera 125 is an example of an image-capturing device.
- the control device 127 can calculate a distance between the stereo camera 125 and a captured target by using an image captured by the right-side camera 1251 and an image captured by the left-side camera 1252 .
- the image captured by the right-side camera 1251 is also referred to as a right-eye image.
- the image captured by the left-side camera 1252 is also referred to as a left-eye image.
- a combination of the images captured by both cameras of the stereo camera 125 is also referred to as a stereo image.
- the stereo camera 125 may be configured of three or more cameras.
- FIG. 3 is a schematic block diagram showing a configuration of the control device according to the first embodiment.
- the control device 127 includes a processor 91 , a main memory 92 , a storage 93 , and an interface 94 .
- a program for controlling the work equipment 110 is stored in the storage 93 .
- a hard disk drive (HDD) and a non-volatile memory are exemplary examples of the storage 93 .
- the storage 93 may be an internal medium directly connected to a bus of the control device 127 or an external medium connected to the control device 127 through the interface 94 or a communication line.
- the storage 93 is an example of a storage unit.
- the processor 91 reads the program from the storage 93 , expands the program into the main memory 92 , and executes processing according to the program.
- the processor 91 secures a storage area in the main memory 92 according to the program.
- the main memory 92 is an example of the storage unit.
- the interface 94 is connected to the work equipment position detector 122 , the position azimuth direction calculator 123 , the inclination detector 124 , the stereo camera 125 , the display device 128 , and another peripheral device to transmit and receive signals.
- the control device 127 displays the guidance information representing a relative position between the work machine and the loading/unloading target of the transporting material on the display device 128 . Accordingly, the operator can recognize the loading/unloading target by referring to the guidance information even in a case where it is difficult to visually recognize the loading/unloading target.
- the processor 91 is provided with a data acquisition unit 1701 , an area-specifying unit 1702 , a three-dimensional data generation unit 1703 , a loading/unloading target-specifying unit 1704 , a model-fitting unit 1705 , a work equipment position-specifying unit 1706 , a guidance information generation unit 1707 , a display control unit 1708 , and a learning unit 1709 .
- the storage 93 stores a camera parameter CP, a segmentation model M, and a target model D.
- the camera parameter CP is information indicating a positional relationship between the swing body 120 and the right-side camera 1251 and a positional relationship between the swing body 120 and the left-side camera 1252 .
- the target model D is three-dimensional data representing a shape of the dump body 210 .
- the program may be for realizing some of the functions exerted by the control device 127 .
- the program may exert the function by a combination with another program already stored in the storage 93 or a combination with another program installed in another device.
- the control device 127 may include a custom large-scale integrated circuit (LSI) such as a programmable logic device (PLD) in addition to or instead of the above configuration.
- LSI large-scale integrated circuit
- a programmable array logic (PAL), a generic array logic (GAL), a complex programmable logic device (CPLD), and a field-programmable gate array (FPGA) are exemplary examples of the PLD.
- some or all of the functions realized by the processor may be realized by the integrated circuit.
- the data acquisition unit 1701 acquires data from the work equipment position detector 122 , the position azimuth direction calculator 123 , the inclination detector 124 , and the stereo camera 125 through the interface 94 . That is, the data acquisition unit 1701 acquires the angle of the work equipment 110 , a position, azimuth direction, and inclination of the swing body 120 , and the stereo image.
- the data acquisition unit 1701 is an example of an image acquisition unit.
- the position, azimuth direction, and inclination of the swing body 120 are also referred to as a posture of the swing body 120 . Since the stereo camera 125 is provided on the swing body 120 , the position, azimuth direction, and inclination of the swing body 120 represent a posture of the stereo camera 125 .
- FIG. 4 is a diagram showing an example of image segmentation according to the first embodiment.
- the area-specifying unit 1702 inputs the right-eye image of the stereo image acquired by the data acquisition unit 1701 into the segmentation model M stored in the storage 93 to divide the right-eye image into a plurality of areas corresponding to a plurality of known objects. For example, as shown in FIG. 4 , the area-specifying unit 1702 generates a segment image in which the right-eye image which is an input image is divided into the plurality of areas corresponding to five objects of a vehicle body of the dump truck 200 , the dump body 210 , the transporting material L, the work equipment 110 , and the ground.
- the plurality of known objects include at least the dump body 210 which is a loading target.
- a value of each pixel of the segment image takes a value representing a type of the object shown in the pixel of the input image corresponding to the pixel. That is, the segment image is divided into the plurality of areas (segments) formed of a plurality of pixels having the same value.
- the area-specifying unit 1702 specifies an area in which the dump body is shown in the stereo image acquired by the data acquisition unit 1701 based on the segment image of the segmentation model M.
- the segmentation model M includes a neural network 140 shown in FIG. 5 .
- FIG. 5 is a diagram showing an example of a configuration of a neural network.
- the segmentation model M is realized by, for example, a learned model of a deep neural network (DNN).
- the learned model is configured of a combination of a learning model and a learned parameter.
- the neural network 140 includes an input layer 141 , one or more intermediate layers 142 (hidden layers), and an output layer 143 .
- Each of the layers 141 , 142 , and 143 includes one or more neurons.
- the number of neurons in the intermediate layer 142 can be set as appropriate.
- the output layer 143 can be set as appropriate according to a size of an output image.
- Neurons in layers adjacent to each other are coupled to each other, and a weight (coupling load) is set for each coupling.
- the number of coupled neurons may be set as appropriate.
- a threshold value is set for each neuron.
- An output value of each neuron is determined by whether or not a sum of products of an input value and the weight for each neuron exceeds the threshold value.
- An image is input to the input layer 141 .
- the image input to the input layer 141 is an image showing an object corresponding to at least one of a plurality of areas (a plurality of objects).
- the output layer 143 a segment image in which a value of each pixel takes a value representing a type of an object that is shown in a corresponding pixel of the image input to the input layer 141 and corresponds to the pixel is input. That is, the segmentation model M is a learned model trained to output a segment image divided into areas for each object shown in an image when the image is input.
- the segmentation model M is trained by using a data set for learning in which the image captured at the loading site is used as learning data and the segment image in which the pixel value is divided for each of the vehicle body of the dump truck 200 , the dump body 210 , the transporting material L, the work equipment 110 , and the ground is used as supervised data.
- the “learning data” refers to data input to the input layer during the training of the learning model.
- the “supervised data” is data which is a correct answer for comparison with the value of the output layer of the neural network 140 .
- the “data set for learning” refers to a combination of the learning data and the supervised data.
- the learned parameter of the segmentation model M obtained by learning is stored in the storage 93 .
- the learned parameter includes, for example, the number of layers of the neural network 140 , the number of neurons in each layer, a coupling relationship between the neurons, the weight of the coupling between the neurons, and the threshold value of each neuron.
- a DNN configuration generally used for image segmentation is used as the configuration of the neural network 140 of the segmentation model M.
- the segmentation model M may be a transfer-learned version of an existing learned segmentation model using the above data set for learning.
- the segmentation model M according to another embodiment may be trained by unsupervised learning or reinforcement learning.
- the three-dimensional data generation unit 1703 extracts, based on an area in which the dump body 210 specified by the area-specifying unit 1702 is shown, a partial right-eye image and a partial left-eye image configured of pixels related to the area from each of the right-eye image and the left-eye image of the stereo image.
- the three-dimensional data generation unit 1703 may extract the partial right-eye image and the partial left-eye image not only from the area classified as the dump body 210 but also from an area including the vicinity of the area classified into the dump body 210 .
- the three-dimensional data generation unit 1703 may extract an area in which a rectangle circumscribing the area classified as the dump body 210 is expanded vertically and horizontally by a predetermined number of pixels as the partial right-eye image and the partial left-eye image.
- the area including the vicinity of the area classified as the dump body 210 it is possible to prevent the area in which the dump body 210 is shown from being missing even in a case where a part of the dump body is classified as another object due to a segmentation error.
- the three-dimensional data generation unit 1703 generates point group data indicating a position of the area in which the dump body 210 is shown in the site coordinate system based on the partial right-eye image and the partial left-eye image, the camera parameter CP stored in the storage 93 , and the position, azimuth direction, and inclination of the swing body 120 acquired by the data acquisition unit 1701 .
- the three-dimensional data generation unit 1703 generates the point group data by the following method. First, the three-dimensional data generation unit 1703 generates the point group data related to the vehicle body coordinate system by triangulation based on the partial right-eye image, the partial left-eye image, and the camera parameter CP.
- the three-dimensional data generation unit 1703 converts a position of each point in the vehicle body coordinate system into the position of each point in the site coordinate system based on the position, azimuth direction, and inclination of the swing body 120 .
- the point group data is an example of three-dimensional data representing a three-dimensional shape of the subject of the captured image.
- a depth image, a polygon, a CAD model, and the like may be used as the three-dimensional data.
- the loading/unloading target-specifying unit 1704 specifies a three-dimensional feature amount (3D point group feature amount) for each point of the point group data generated by the three-dimensional data generation unit 1703 .
- a normal feature, a signature of histograms of orientations (SHOT) feature, and a point feature histograms (PFH) feature are exemplary examples of the three-dimensional feature amount.
- the normal feature of a certain point (point of interest) included in the point group can be obtained by obtaining an eigenvector related to a minimum eigenvalue of a covariance matrix of the point group around the point of interest.
- the loading/unloading target-specifying unit 1704 groups the point group data for each surface of the dump body 210 based on the feature amount.
- the loading/unloading target-specifying unit 1704 divides the point group data into a group corresponding to a front panel, a group corresponding to a side gate, a group corresponding to a tailgate, and a group corresponding to a bottom surface, based on the feature amount.
- the division method of the point group data described here is an example, and the point group data may be divided into groups other than the above four groups.
- the loading/unloading target-specifying unit 1704 may not perform the grouping into the group corresponding to the bottom surface.
- the loading/unloading target-specifying unit 1704 is an example of a dump body surface-specifying unit.
- the model-fitting unit 1705 disposes the target model D in a virtual space represented by the site coordinate system based on the position of the point group data grouped for each surface to specify a position, azimuth direction, and inclination of the dump body 210 in the site coordinate system. For example, the model-fitting unit 1705 disposes the target model D such that a typical normal feature of the group of the point group data and the normal feature of each surface of the dump body 210 match within a predetermined error range.
- the model-fitting unit 1705 may specify the position of each surface in the vehicle body coordinate system or a camera coordinate system.
- the model-fitting unit 1705 is an example of a posture-specifying unit.
- the model-fitting unit 1705 is an example of a posture-specifying unit.
- the work equipment position-specifying unit 1706 specifies positions of the boom 111 , the arm 112 , and the bucket 113 in the site coordinate system based on the angle of the work equipment 110 and the position, azimuth direction, and inclination of the swing body 120 acquired by the data acquisition unit 1701 .
- the guidance information generation unit 1707 generates the guidance information indicating the positional relationship between the dump body 210 and the bucket 113 based on the position of the dump body 210 specified by the model-fitting unit 1705 , the positions of the boom 111 , the arm 112 , and the bucket 113 specified by the work equipment position-specifying unit 1706 , and the position, azimuth direction, and inclination of the swing body 120 acquired by the data acquisition unit 1701 .
- FIG. 6 is an example of the guidance information.
- the guidance information generation unit 1707 disposes the three-dimensional model of the hydraulic excavator 100 in the virtual space based on the positions of the boom 111 , the arm 112 , and the bucket 113 specified by the work equipment position-specifying unit 1706 and the position, azimuth direction, and inclination of the swing body 120 .
- the guidance information generation unit 1707 disposes the three-dimensional model of the hydraulic excavator 100 such that a difference between the three-dimensional feature amount of the three-dimensional model of the dump truck 200 and the three-dimensional feature amount of the dump body 210 specified by the model-fitting unit 1705 is minimized.
- the guidance information generation unit 1707 renders the hydraulic excavator 100 and the dump truck 200 disposed in the virtual space from a random viewpoint to generate the guidance information representing the positional relationship between the bucket 113 of the hydraulic excavator 100 and the dump body 210 of the dump truck 200 .
- the guidance information according to another embodiment may not graphically draw the positional relationship between the bucket 113 and the dump body 210 .
- the guidance information according to another embodiment may be information indicating whether or not the bucket 113 is located in the outer frame of the dump body 210 in plan view from above.
- the guidance information according to another embodiment may be the display of a target excavation position of the bucket 113 or the guidance information for the excavation position of the bucket 113 (for example, information or display indicating a difference between a current position of the bucket 113 and the target excavation position).
- the guidance information may be information or display indicating how much operation is required to reach the target excavation position. At least only the dump body 210 and the bucket 113 may be displayed.
- the display control unit 1708 outputs a display signal for displaying the guidance information to the display device 128 .
- the learning unit 1709 performs learning processing of the segmentation model M.
- the learning unit 1709 may be provided in a device separate from the control device 127 . In this case, the learned model learned in the separate device is recorded in the storage 93 .
- FIG. 7 is a flowchart showing a display method of the guidance information by the control device according to the first embodiment.
- the data acquisition unit 1701 acquires the angle of the work equipment 110 from the work equipment position detector 122 through the interface 94 , acquires the position and azimuth direction of the swing body 120 from the position azimuth direction calculator 123 , acquires the inclination of the swing body 120 from the inclination detector 124 , and acquires the stereo image from the stereo camera 125 (step S 1 ).
- the data acquisition unit 1701 may not acquire the angle of the work equipment 110 , and the position, azimuth direction, and inclination of the swing body 120 .
- the area-specifying unit 1702 inputs the right-eye image of the stereo image acquired by the data acquisition unit 1701 into the segmentation model M stored in the storage 93 to obtain the segment image divided into the plurality of areas for each shown object (step S 2 ).
- the area-specifying unit 1702 may obtain the segment image from the left-eye image.
- the area-specifying unit 1702 specifies an area in which the dump body is shown in the stereo image acquired by the data acquisition unit 1701 based on the segment image (step S 3 ).
- the three-dimensional data generation unit 1703 extracts the partial right-eye image and the partial left-eye image related to the area, specified in step S 3 , in which the dump body 210 is shown from each of the right-eye image and the left-eye image of the stereo image (step S 4 ).
- the three-dimensional data generation unit 1703 generates the point group data related to the vehicle body coordinate system by triangulation based on the partial right-eye image, the partial left-eye image, and the camera parameter CP (step S 5 ).
- the three-dimensional data generation unit 1703 converts the position of each point of the point group data in the vehicle body coordinate system to the position of each point thereof in the site coordinate system based on the position, azimuth direction, and inclination of the swing body 120 acquired in step S 1 (step S 6 ). That is, the three-dimensional data generation unit 1703 moves the point group data in parallel based on the position of the swing body 120 acquired in step S 1 and rotates the point group data based on the azimuth direction, and inclination thereof.
- the coordinate conversion from the vehicle body coordinate system to the site coordinate system may be performed after step S 7 or step S 8 , which will be described below.
- the loading/unloading target-specifying unit 1704 specifies the three-dimensional feature amount for each point of the point group data obtained in step S 6 , and groups the point group data for each surface of the dump body 210 based on the three-dimensional feature amount (step S 7 ). Accordingly, the loading/unloading target-specifying unit 1704 can specify each surface of the dump body 210 from the point group data.
- the model-fitting unit 1705 disposes the target model D in the virtual space represented by the site coordinate system such that a difference between a three-dimensional feature amount of the point group data related to each group and a three-dimensional feature amount of the target model D stored in the storage 93 is minimized to specify the position, azimuth direction, and inclination of the dump body 210 in the site coordinate system (step S 8 ).
- the disposition of the target model D in the virtual space based on the point group data by the model-fitting unit 1705 it is possible to estimate a hidden portion of the dump body 210 even in a case where a part of the dump body 210 or the like is hidden by an obstacle in the stereo image and then it is possible to improve the recognition accuracy of the dump body 210 .
- the model-fitting unit 1705 may fit a certain surface grouped in step S 7 or a specific point of the surface with a corresponding surface or a specific point of the target model D to dispose the target model D in the site coordinate system.
- the control device 127 can specify a position, azimuth direction, and inclination of the dump body 210 from each surface specified in step S 7 without performing the fitting with the target model D.
- control device 127 can specify the position of the dump body 210 from positions of the centers of gravity of the plurality of specified surfaces, specify the azimuth direction of the dump body from a normal direction of the surface corresponding to the front panel or the tailgate, and further specify the inclination of the dump body from the normal direction of the bottom surface of the dump body 210 .
- the control device 127 performs the fitting with the target model D and it is possible to further improve the robustness in specifying the position, azimuth direction, and inclination of the dump body 210 .
- the work equipment position-specifying unit 1706 specifies the positions of the boom 111 , the arm 112 , and the bucket 113 in the site coordinate system based on the angle of the work equipment 110 and the position, azimuth direction, and inclination of the swing body 120 acquired by the data acquisition unit 1701 (step S 9 ).
- the guidance information generation unit 1707 generates the guidance information shown in FIG. 6 based on the position, azimuth direction, and inclination of the dump body 210 specified in step S 8 , the positions of the boom 111 , the arm 112 , and the bucket 113 specified in step S 9 , and the position, azimuth direction, and inclination of the swing body 120 acquired in step S 1 (step S 10 ).
- the display control unit 1708 outputs the display signal for displaying the guidance information to the display device 128 (step S 11 ).
- FIG. 8 is a flowchart showing a learning method of the segmentation model according to the first embodiment.
- the data acquisition unit 1701 acquires the learning data (step S 101 ).
- the learning data in the segmentation model M is an image captured at the loading site or an image generated from 3DCG or CAD data.
- a plurality of learning data used for the learning of the segmentation model M include at least an image showing the dump body 210 .
- the learning data may be acquired from the image captured by the stereo camera 125 .
- the learning data may be acquired from an image captured by another work machine.
- the learning unit 1709 performs the learning of the segmentation model M.
- the learning unit 1709 uses the combination of the learning data acquired in step S 101 and the supervised data which is the segment image divided into the plurality of areas corresponding to the objects shown in the image related to the learning data as the data set for learning to perform the learning of the segmentation model M (step S 102 ).
- the learning unit 1709 uses the learning data as an input to perform calculation processing of the neural network 140 in a forward propagation direction. Accordingly, the learning unit 1709 obtains the output value output from the output layer 143 of the neural network 140 .
- the data set for learning may be stored in the main memory 92 or the storage 93 .
- the learning unit 1709 calculates an error between the output value from the output layer 143 and the supervised data.
- the learning unit 1709 updates the weight of the coupling between the neurons based on the error of the calculated output value.
- the learning unit 1709 inputs the learning data related to an evaluation data set to the segmentation model M to determine whether or not the output value from the segmentation model M matches the supervised data related to the evaluation data set, using the data set for learning that has not been used for the learning of the segmentation model M as the evaluation data set (step S 103 ).
- the number of pixels in which the output value and the supervised data are different is within a predetermined number, it may be determined that the output value and the supervised data match.
- the above processing is repeated until the output value from the segmentation model M matches the supervised data.
- a parameter of the segmentation model M is optimized, and the segmentation model M can be learned.
- the learning unit 1709 records the segmentation model M which is a learned model including the parameter optimized by the learning in the storage 93 (step S 104 ).
- the control device 127 specifies the area in which the dump body 210 is shown and specifies the position of the dump body 210 based on the area, from the captured image showing the dump body 210 which is the loading/unloading target of the transporting material L.
- the control device 127 according to the first embodiment can specify the position of the loading/unloading target regardless of an edge included in the image. Accordingly, the control device 127 according to the first embodiment can improve the robustness of the processing of specifying the position of the loading/unloading target of the transporting material L even in a case where it is difficult to detect the edge due to the distortion of the loading/unloading target or the like.
- the transporting material L When the transporting material L is loaded into the dump body 210 , a surface shape of the dump body 210 as a whole is changed. However, the transporting material L and the dump body 210 can be identified by specifying the area in which the dump body 210 is shown. Therefore, it is possible to improve the robustness of recognition. Depending on the embodiment, the processing of steps S 2 to S 6 and steps S 8 to S 11 among the processing by the control device 127 shown in FIG. 7 may not be performed.
- the control device 127 displays the guidance information. Accordingly, even in a case where it is difficult for the operator to visually recognize the dump body 210 due to a relative position between the hydraulic excavator 100 and the dump truck 200 , the control device 127 allows the operator to recognize the loading/unloading target. It is possible to improve the work efficiency by operating the work equipment 110 using the dump body 210 displayed as the guidance information as a guide.
- the control device 127 specifies the area in which the loading/unloading target is shown in the captured image based on the segmentation model M and the captured image. It is known that the segmentation, which is the segmentation of objects included in an image, can be accurately realized with the learned model related to machine learning. Therefore, according to the first embodiment, the control device 127 can accurately specify a position of a predetermined portion of the dump body 210 by recognizing the dump body 210 using the segmentation model M.
- the present invention is not limited thereto in another embodiment.
- the control device 127 may perform the segmentation of the objects shown in the image by a method other than the machine learning technique.
- the control device 127 generates the point group data using the partial right-eye image and the partial left-eye image related to the specified area. Accordingly, it is possible to reduce an amount of calculation as compared with the case where the point group data is generated using all the pixels of the stereo image. Since the extra areas other than the area specified by the segmentation (for example, areas such as the transporting material L, the work equipment 110 , and the ground) are omitted, the loading/unloading target-specifying unit 1704 can reduce a possibility of noise being mixed in the group of the point group data.
- the control device 127 may generate the point group data using all the pixels of the stereo image and extract a portion related to a specified area from the point group data.
- the three-dimensional data may be generated by using a distant image generated by a laser scanner instead of the stereo image.
- the control device 127 generates the point group data from the captured image and disposes the point group data in the site coordinate system based on the position, azimuth direction, and inclination of the swing body 120 . Accordingly, it is possible to specify the position, azimuth direction, and inclination of the dump body 210 in the site coordinate system even in a case where a position, azimuth direction, and inclination of the stereo camera 125 are changed by the operation of the hydraulic excavator 100 . In another embodiment, in a case where the stereo camera 125 is fixedly installed at the construction site, the point group data may be disposed in the site coordinate system without obtaining the position, azimuth direction, and inclination of the swing body 120 .
- the control device 127 according to the first embodiment generates the guidance information based on the specified position, azimuth direction, and inclination of the dump body 210 and presents the guidance information to the operator. On the contrary, the control device 127 according to the second embodiment controls the loading work of the work equipment based on the specified position, azimuth direction, and inclination of the dump body 210 .
- FIG. 9 is a schematic block diagram showing a configuration of a control device according to the second embodiment.
- the processor 91 according to the second embodiment includes a work machine control unit 1711 instead of the guidance information generation unit 1707 and the display control unit 1708 according to the first embodiment.
- the work machine control unit 1711 controls behavior of the swing body 120 and the work equipment 110 based on the position, azimuth direction, and inclination of the dump body 210 specified by the model-fitting unit 1705 in a case where an automatic loading button provided on the operation device 126 is pressed by the operator.
- FIG. 10 is a flowchart showing a control method of a hydraulic excavator according to the second embodiment.
- the control device 127 specifies the positions of the dump body 210 and the work equipment 110 in the site coordinate system by the same processing as in steps S 1 to S 9 of the first embodiment.
- the work machine control unit 1711 generates and outputs control signals for the swing body 120 and the work equipment 110 based on the positions of the dump body 210 and the work equipment 110 in the site coordinate system (step S 31 ).
- the work machine control unit 1711 In a case where a height of the bucket 113 is lower than a height of the dump body 210 , the work machine control unit 1711 generates the control signal for raising the boom 111 . In a case where the height of the bucket 113 is equal to or higher than the height of the dump body 210 , the work machine control unit 1711 stops the output of the control signal for raising the boom 111 and starts the output of the control signal for swing the swing body 120 .
- the work machine control unit 1711 stops the output of the control signal for swing the swing body 120 and generates the control signal for dumping the bucket 113 .
- control device 127 can automatically control the hydraulic excavator 100 based on the captured image.
- the control device 127 causes the display device 128 to display a dump body map indicating a distribution of an amount of the transporting material L in the dump body 210 . Accordingly, the control device 127 causes the operator to recognize the loading/unloading position of the transporting material L for loading the transporting material L into the dump body 210 in a well-balanced manner.
- FIG. 11 is a schematic block diagram showing a configuration of a control device according to the third embodiment.
- the control device 127 according to the third embodiment includes a distribution-specifying unit 1721 instead of the work equipment position-specifying unit 1706 according to the first embodiment.
- the area-specifying unit 1702 , the three-dimensional data generation unit 1703 , and the guidance information generation unit 1707 according to the third embodiment have different processing from that of the first embodiment.
- the area-specifying unit 1702 specifies an area in which the transporting material L is shown in addition to the area in which the dump body 210 is shown in the stereo image.
- the three-dimensional data generation unit 1703 generates dump body point group data, which is three-dimensional data related to the area in which the dump body 210 is shown, and transporting material point group data, which is three-dimensional data related to the area in which the transporting material L is shown.
- the three-dimensional data generation unit 1703 may generate the transporting material point group data not only from the area classified into the transporting material L but also from the area including the vicinity of the area classified into the transporting material L.
- the three-dimensional data generation unit 1703 may generate the transporting material point group data based on the area in which the rectangle circumscribing the area classified into the transporting material L is expanded vertically and horizontally by a predetermined number of pixels. By extracting the area including the vicinity of the area classified as the transporting material L, it is possible to prevent the area in which the transporting material L is shown from being missing even in a case where a part of the transporting material L is classified as another object due to a segmentation error.
- the distribution-specifying unit 1721 generates the dump body map indicating the distribution of the amount of the transporting material L in the dump body 210 based on a three-dimensional position of the bottom surface of the dump body 210 specified by the model-fitting unit 1705 and the transporting material point group data generated by the three-dimensional data generation unit 1703 .
- the dump body map is, for example, an elevation map of the transporting material L with reference to the bottom surface of the dump body 210 .
- the guidance information generation unit 1707 generates the guidance information from the dump body map generated by the distribution-specifying unit 1721 .
- FIG. 12 is an example of guidance information according to the third embodiment.
- the guidance information generation unit 1707 generates the guidance information for displaying a two-dimensional heat map representing a distribution of a height from the bottom surface of the dump body 210 to the surface of the transporting material L.
- Granularity of vertical and horizontal divisions in the heat map shown in FIG. 12 is an example and is not limited thereto in another embodiment.
- the heat map according to another embodiment may represent, for example, a ratio of a height of the transporting material L to a height related to an upper limit of the loading of the dump body 210 .
- FIG. 13 is a flowchart showing a display method of the guidance information by the control device according to the third embodiment.
- the data acquisition unit 1701 acquires the position and azimuth direction of the swing body 120 from the position azimuth direction calculator 123 through the interface 94 , acquires the inclination of the swing body 120 from the inclination detector 124 , and acquires the stereo image from the stereo camera 125 (step S 51 ). In another embodiment, the data acquisition unit 1701 may not acquire the angle of the work equipment 110 , and the position, azimuth direction, and inclination of the swing body 120 .
- the area-specifying unit 1702 inputs the stereo image acquired by the data acquisition unit 1701 into the segmentation model M stored in the storage 93 to obtain the segment image divided into the plurality of areas for each shown object (step S 52 ).
- the area-specifying unit 1702 specifies the area in which the dump body 210 is shown and the area in which the transporting material L is shown in the stereo image acquired by the data acquisition unit 1701 , based on the segment image (step S 53 ).
- the three-dimensional data generation unit 1703 extracts the partial right-eye image and the partial left-eye image related to the area, specified in step S 3 , in which the dump body 210 is shown, and the partial right-eye image and the partial left-eye image related to the area in which the transporting material L is shown, from each of the right-eye image and the left-eye image of the stereo image (step S 54 ).
- the three-dimensional data generation unit 1703 generates the dump body point group data related to the vehicle body coordinate system by triangulation based on the partial right-eye image and the partial left-eye image related to the area in which the dump body 210 is shown and the camera parameter CP.
- the three-dimensional data generation unit 1703 generates the transporting material point group data related to the vehicle body coordinate system by triangulation based on the partial right-eye image and the partial left-eye image related to the area in which the transporting material L is shown and the camera parameter CP (step S 55 ).
- the three-dimensional data generation unit 1703 converts the position of each point of the dump body point group data and the transporting material point group data in the vehicle body coordinate system to the position of each point thereof in the site coordinate system, based on the position, azimuth direction, and inclination of the swing body 120 (step S 56 ).
- the loading/unloading target-specifying unit 1704 specifies the three-dimensional feature amount for each point of the dump body point group data obtained in step S 56 and groups the dump body point group data for each surface of the dump body 210 based on the three-dimensional feature amount (step S 57 ).
- the model-fitting unit 1705 disposes the target model D in the virtual space represented by the site coordinate system such that a difference between a three-dimensional feature amount of the dump body point group data related to each group and the three-dimensional feature amount of the target model D stored in the storage 93 is minimized to specify the position, azimuth direction, and inclination of the dump body 210 in the site coordinate system (step S 58 ).
- the distribution-specifying unit 1721 generates the dump body map which is the elevation map representing a height related to the vertically upward direction with reference to the bottom surface of the dump body 210 based on the transporting material point group data generated in step S 55 and the target model D disposed in step S 58 (step S 59 ).
- the dump body map may include a grid without the height data.
- the distribution-specifying unit 1721 can generate the dump body map by obtaining the elevation map with an XY plane as the reference height and with the Z-axis direction as the height direction.
- the guidance information generation unit 1707 generates the guidance information shown in FIG. 12 based on the dump body map (step S 60 ).
- the display control unit 1708 outputs the display signal for displaying the guidance information to the display device 128 (step S 61 ).
- the control device 127 specifies the three-dimensional positions of the surface of the transporting material L and the bottom surface of the dump body 210 based on the captured image to generate the dump body map indicating the distribution of the amount of the transporting material L in the dump body 210 based on the three-dimensional positions. Accordingly, the control device 127 can specify the distribution of the transporting material L in the dump body 210 . The operator can recognize the loading/unloading position of the transporting material L for loading the transporting material L in the dump body 210 in a well-balanced manner by recognizing the distribution of the transporting material L in the dump body 210 .
- the control device 127 generates the guidance information based on the dump body map, but the present invention is not limited thereto.
- the control device 127 may control the hydraulic excavator 100 as in the second embodiment based on the dump body map.
- the work machine control unit 1711 generates the control signal for raising the boom 111 .
- the work machine control unit 1711 stops the output of the control signal for raising the boom 111 and starts the output of the control signal for swing the swing body 120 .
- the work machine control unit 1711 stops the output of the control signal for swing the swing body 120 and generates the control signal for dumping the bucket 113 .
- the control device 127 according to the third embodiment matches the target model D with the dump body point group data to specify the three-dimensional position of the bottom surface of the dump body 210 .
- the control device 127 according to the fourth embodiment specifies the three-dimensional position of the bottom surface of the dump body 210 without using the target model D. That is, the control device 127 according to the fourth embodiment may not store the target model D in the storage 93 .
- the model-fitting unit 1705 distorts a rectangular parallelepiped polygon without a top surface to best match the dump body point group data based on the dump body point group data.
- the model-fitting unit 1705 specifies a position of a bottom surface of the matched rectangular parallelepiped polygon as the position of the bottom surface of the dump body 210 .
- the fourth embodiment it is possible to specify the position of the bottom surface of the dump body 210 without using the target model D of the dump body 210 . According to the fourth embodiment, it is possible to specify the position of the bottom surface of the dump body 210 even though the dump body point group data contains noise.
- FIG. 14 is a diagram showing an example of image segmentation according to a fifth embodiment.
- the area in which the dump body 210 is shown is represented by one segment.
- the control device 127 according to the fifth embodiment may not include the loading/unloading target-specifying unit 1704 in the configuration of the first embodiment.
- the operations of the area-specifying unit 1702 , the three-dimensional data generation unit 1703 , and the model-fitting unit 1705 of the control device 127 according to the fifth embodiment are different.
- the area-specifying unit 1702 inputs the right-eye image of the stereo image acquired by the data acquisition unit 1701 into the segmentation model M stored in the storage 93 to divide the right-eye image into the plurality of areas corresponding to the plurality of known objects.
- the plurality of known objects include at least the guard frame, the front panel, the side gate, and the tailgate.
- the three-dimensional data generation unit 1703 generates the point group data related to each area of the guard frame, the front panel, the side gate, and the tailgate from the stereo image.
- the model-fitting unit 1705 disposes the target model D in the virtual space represented by the site coordinate system based on the position of each piece of point group data of the guard frame, the front panel, the side gate, and the tailgate to specify the position, azimuth direction, and inclination of the dump body 210 in the site coordinate system.
- FIG. 15 is a flowchart showing a display method of the guidance information by a control device according to the fifth embodiment.
- the data acquisition unit 1701 acquires the angle of the work equipment 110 from the work equipment position detector 122 through the interface 94 , acquires the position and azimuth direction of the swing body 120 from the position azimuth direction calculator 123 , acquires the inclination of the swing body 120 from the inclination detector 124 , and acquires the stereo image from the stereo camera 125 (step S 71 ).
- the data acquisition unit 1701 may not acquire the angle of the work equipment 110 , and the position, azimuth direction, and inclination of the swing body 120 .
- the area-specifying unit 1702 inputs the stereo image acquired by the data acquisition unit 1701 into the segmentation model M stored in the storage 93 to obtain the segment image divided into the plurality of areas for each shown object (step S 72 ).
- the area-specifying unit 1702 specifies the area for each surface of the dump body 210 in the stereo image acquired by the data acquisition unit 1701 based on the segment image (step S 73 ). That is, the area-specifying unit 1702 specifies each of the area in which the guard frame is shown, the area in which the front panel is shown, the area in which the side gate is shown, and the area in which the tailgate is shown.
- the three-dimensional data generation unit 1703 extracts the partial right-eye image and the partial left-eye image related to each surface specified in step S 73 from each of the right-eye image and the left-eye image of the stereo image (step S 74 ).
- the three-dimensional data generation unit 1703 generates the point group data of each surface related to the vehicle body coordinate system by triangulation based on the partial right-eye image, the partial left-eye image, and the camera parameter CP (step S 75 ).
- the three-dimensional data generation unit 1703 converts the position of each point of the point group data of each surface in the vehicle body coordinate system to the position of each point thereof in the site coordinate system based on the position, azimuth direction, and inclination of the swing body 120 (step S 76 ).
- the point group data is already divided into the group for each surface without specifying the three-dimensional feature amount.
- the model-fitting unit 1705 disposes the target model D in the virtual space represented by the site coordinate system such that a difference between the position of the point group data related to each group and the position of each surface of the target model D stored in the storage 93 is minimized to specify the position, azimuth direction, and inclination of the dump body 210 in the site coordinate system (step S 77 ).
- the work equipment position-specifying unit 1706 specifies the positions of the boom 111 , the arm 112 , and the bucket 113 in the site coordinate system based on the angle of the work equipment 110 and the position, azimuth direction, and inclination of the swing body 120 acquired by the data acquisition unit 1701 (step S 78 ).
- the guidance information generation unit 1707 generates the guidance information based on the position, azimuth direction, and inclination of the dump body 210 specified in step S 77 , the positions of the boom 111 , the arm 112 , and the bucket 113 specified in step S 78 , and the position, azimuth direction, and inclination of the swing body 120 acquired in step S 71 (step S 79 ).
- the display control unit 1708 outputs the display signal for displaying the guidance information to the display device 128 (step S 80 ).
- the segmentation model M divides each surface of the dump body 210 into the different areas. Accordingly, the control device 127 can match the three-dimensional data with the three-dimensional model without calculating the three-dimensional feature amount.
- control device 127 generates the three-dimensional data from the stereo image based on triangulation, but the present invention is not limited thereto.
- control device 127 may generate the three-dimensional data using a learned model that outputs the depth image by inputting the stereo image.
- segmentation model M learned such that the stereo image is input and the depth image divided into areas for each object is output may be employed.
- control device 127 specifies the position, azimuth direction, and inclination of the dump body 210 , but is not limited thereto in another embodiment.
- the control device 127 according to another embodiment may specify the position of the dump body 210 and may not specify the azimuth direction, and inclination of the dump body 210 .
- the control device 127 according to another embodiment may specify the position and azimuth direction of the dump body 210 and may not specify the inclination of the dump body 210 .
- the control device 127 is mounted on the hydraulic excavator 100 , but the present invention is not limited thereto.
- the control device 127 according to another embodiment may be provided in a remote server device.
- the control device 127 may be realized from a plurality of computers. In this case, a part of the configuration of the control device 127 may be provided in the remote server device. That is, the control device 127 may be implemented as an image processing system formed of a plurality of devices.
- the hydraulic excavator 100 may be a fully autonomous type or a partially autonomous type hydraulic excavator.
- the display signal for displaying the guidance information may be transmitted to a remote cab for remotely controlling the work machine.
- the loading/unloading target according to the above embodiment is the dump body 210 of the dump truck 200 , but the present invention is not limited thereto.
- the loading/unloading target according to another embodiment may be another loading/unloading target such as a hopper.
- the operations described using the flowchart in the above embodiment are not limited to those executed in the above order.
- the coordinate system of the point group data is converted to the site coordinate system in step S 6 , and then the surface grouping in step S 7 and the posture specification of the dump body in step S 8 are performed.
- the present invention is not limited thereto.
- the coordinate conversion may be performed after the surface grouping in step S 7 or the posture specification of the dump body in step S 8 .
- the control device of the hydraulic excavator 100 specifies the position of the dump body 210 , but the present invention is not limited thereto.
- a control device of another work machine such as a mobile crane, a control device of a work machine such as a dump truck, a wheel loader, or a bulldozer, or a monitoring device, which is installed in the site, may specify the position of the dump body 210 .
- the image processing system can improve the robustness of the processing of specifying the loading/unloading target.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- Artificial Intelligence (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
- Operation Control Of Excavators (AREA)
Abstract
Description
-
-
Patent Document 1 - Japanese Unexamined Patent Application, First Publication No. 2000-192514
-
Claims (7)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2019017400A JP7365122B2 (en) | 2019-02-01 | 2019-02-01 | Image processing system and image processing method |
JP2019-017400 | 2019-02-01 | ||
PCT/JP2020/002554 WO2020158611A1 (en) | 2019-02-01 | 2020-01-24 | Image processing system, image processing method, prelearned model generation method, and data set for learning |
Publications (2)
Publication Number | Publication Date |
---|---|
US20220101552A1 US20220101552A1 (en) | 2022-03-31 |
US12094151B2 true US12094151B2 (en) | 2024-09-17 |
Family
ID=71841008
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/422,029 Active 2041-03-10 US12094151B2 (en) | 2019-02-01 | 2020-01-24 | Image processing system, image processing method, learned model generation method, and data set for learning |
Country Status (6)
Country | Link |
---|---|
US (1) | US12094151B2 (en) |
JP (1) | JP7365122B2 (en) |
KR (1) | KR20210110671A (en) |
CN (1) | CN113366533A (en) |
DE (1) | DE112020000310T5 (en) |
WO (1) | WO2020158611A1 (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6840645B2 (en) * | 2017-09-08 | 2021-03-10 | 株式会社小松製作所 | Construction management equipment and construction management method |
JP7088691B2 (en) * | 2018-02-28 | 2022-06-21 | 株式会社小松製作所 | Loading machine control, control method and remote control system |
JP7283332B2 (en) * | 2019-09-26 | 2023-05-30 | コベルコ建機株式会社 | Container measurement system |
US11624171B2 (en) * | 2020-07-31 | 2023-04-11 | Baidu Usa Llc | Engineering machinery equipment, and method, system, and storage medium for operation trajectory planning thereof |
JP7484630B2 (en) * | 2020-09-29 | 2024-05-16 | コベルコ建機株式会社 | Loading point determination system |
CN112348781A (en) * | 2020-10-26 | 2021-02-09 | 广东博智林机器人有限公司 | Method, device and equipment for detecting height of reference plane and storage medium |
JP2022077328A (en) * | 2020-11-11 | 2022-05-23 | 日本電気株式会社 | Control device, control system, and control method |
KR20230105967A (en) * | 2022-01-05 | 2023-07-12 | (주)서울로보틱스 | Method and apparatus for monitoring work site |
JP2023152458A (en) | 2022-04-04 | 2023-10-17 | 日立建機株式会社 | Work machine |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11211438A (en) | 1998-01-22 | 1999-08-06 | Komatsu Ltd | Load carrying platform load volume measuring device |
JP2000192514A (en) | 1998-12-28 | 2000-07-11 | Hitachi Constr Mach Co Ltd | Automatically operating construction machine and operating method thereof |
WO2008018398A1 (en) | 2006-08-10 | 2008-02-14 | Nec Corporation | Object region extracting device |
US20140184643A1 (en) * | 2012-12-27 | 2014-07-03 | Caterpillar Inc. | Augmented Reality Worksite |
US20150352994A1 (en) * | 2013-01-18 | 2015-12-10 | Hino Motors, Ltd. | Bed height adjustment system for vehicle |
WO2017042873A1 (en) | 2015-09-08 | 2017-03-16 | 株式会社日立製作所 | Remote operation system and operation assistance system |
US20170089042A1 (en) * | 2015-09-30 | 2017-03-30 | Komatsu Ltd. | Periphery Monitoring Device for Crawler-Type Working Machine |
US20170120823A1 (en) * | 2011-06-07 | 2017-05-04 | Komatsu Ltd. | Dump truck |
US20170328032A1 (en) | 2014-12-12 | 2017-11-16 | Hitachi, Ltd. | Volume Estimation Device and Work Machine Using Same |
US20180068564A1 (en) | 2016-09-05 | 2018-03-08 | Panasonic Intellectual Property Corporation Of America | Parking position identification method, parking position learning method, parking position identification system, parking position learning device, and non-transitory recording medium for recording program |
JP2018146347A (en) | 2017-03-03 | 2018-09-20 | 株式会社キーエンス | Image processing device, image processing method, and computer program |
US20180285698A1 (en) | 2017-03-31 | 2018-10-04 | Fujitsu Limited | Image processing apparatus, image processing method, and image processing program medium |
US20190026914A1 (en) * | 2017-07-24 | 2019-01-24 | Deere & Company | Estimating a volume of contents in a container of a work vehicle |
WO2019016870A1 (en) | 2017-07-18 | 2019-01-24 | 株式会社ソニー・インタラクティブエンタテインメント | Image recognition device, image recognition method, and program |
JP2019016294A (en) | 2017-07-10 | 2019-01-31 | 株式会社リコー | Information processor, information processing method, information processing program, and system |
US20190048560A1 (en) * | 2016-04-21 | 2019-02-14 | Sumitomo(S.H.I.) Construction Machinery Co., Ltd. | Display device for shovel |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6633575B2 (en) | 2017-07-11 | 2020-01-22 | 株式会社三共 | Gaming machine |
-
2019
- 2019-02-01 JP JP2019017400A patent/JP7365122B2/en active Active
-
2020
- 2020-01-24 CN CN202080010943.6A patent/CN113366533A/en active Pending
- 2020-01-24 KR KR1020217024275A patent/KR20210110671A/en not_active Application Discontinuation
- 2020-01-24 US US17/422,029 patent/US12094151B2/en active Active
- 2020-01-24 WO PCT/JP2020/002554 patent/WO2020158611A1/en active Application Filing
- 2020-01-24 DE DE112020000310.8T patent/DE112020000310T5/en active Pending
Patent Citations (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11211438A (en) | 1998-01-22 | 1999-08-06 | Komatsu Ltd | Load carrying platform load volume measuring device |
JP2000192514A (en) | 1998-12-28 | 2000-07-11 | Hitachi Constr Mach Co Ltd | Automatically operating construction machine and operating method thereof |
WO2008018398A1 (en) | 2006-08-10 | 2008-02-14 | Nec Corporation | Object region extracting device |
US20090304271A1 (en) | 2006-08-10 | 2009-12-10 | Yusuke Takahashi | Object region extracting device |
US20170120823A1 (en) * | 2011-06-07 | 2017-05-04 | Komatsu Ltd. | Dump truck |
US20140184643A1 (en) * | 2012-12-27 | 2014-07-03 | Caterpillar Inc. | Augmented Reality Worksite |
US20150352994A1 (en) * | 2013-01-18 | 2015-12-10 | Hino Motors, Ltd. | Bed height adjustment system for vehicle |
US20170328032A1 (en) | 2014-12-12 | 2017-11-16 | Hitachi, Ltd. | Volume Estimation Device and Work Machine Using Same |
WO2017042873A1 (en) | 2015-09-08 | 2017-03-16 | 株式会社日立製作所 | Remote operation system and operation assistance system |
US20170089042A1 (en) * | 2015-09-30 | 2017-03-30 | Komatsu Ltd. | Periphery Monitoring Device for Crawler-Type Working Machine |
US20190048560A1 (en) * | 2016-04-21 | 2019-02-14 | Sumitomo(S.H.I.) Construction Machinery Co., Ltd. | Display device for shovel |
US20180068564A1 (en) | 2016-09-05 | 2018-03-08 | Panasonic Intellectual Property Corporation Of America | Parking position identification method, parking position learning method, parking position identification system, parking position learning device, and non-transitory recording medium for recording program |
JP2018041176A (en) | 2016-09-05 | 2018-03-15 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | Parking position identifying method, parking position learning method, parking position identification system, parking position learning device and program |
JP2018146347A (en) | 2017-03-03 | 2018-09-20 | 株式会社キーエンス | Image processing device, image processing method, and computer program |
US20180285698A1 (en) | 2017-03-31 | 2018-10-04 | Fujitsu Limited | Image processing apparatus, image processing method, and image processing program medium |
JP2018173814A (en) | 2017-03-31 | 2018-11-08 | 富士通株式会社 | Image processing device, image processing method, image processing program and teacher data creating method |
JP2019016294A (en) | 2017-07-10 | 2019-01-31 | 株式会社リコー | Information processor, information processing method, information processing program, and system |
WO2019016870A1 (en) | 2017-07-18 | 2019-01-24 | 株式会社ソニー・インタラクティブエンタテインメント | Image recognition device, image recognition method, and program |
US20200111215A1 (en) * | 2017-07-18 | 2020-04-09 | Sony Interactive Entertainment Inc. | Image recognition apparatus, image recognition method, and program |
US20190026914A1 (en) * | 2017-07-24 | 2019-01-24 | Deere & Company | Estimating a volume of contents in a container of a work vehicle |
Non-Patent Citations (1)
Title |
---|
Rosebrock, "Semantic segmentation with OpenCV and deep learning," Sep. 3, 2018, retrieved from URL <https://pyimagesearch.com/2018/09/03/semantic-segmentation-with-opencv-and-deep-learning/>, 31 pages. |
Also Published As
Publication number | Publication date |
---|---|
US20220101552A1 (en) | 2022-03-31 |
WO2020158611A1 (en) | 2020-08-06 |
CN113366533A (en) | 2021-09-07 |
KR20210110671A (en) | 2021-09-08 |
JP7365122B2 (en) | 2023-10-19 |
DE112020000310T5 (en) | 2021-11-04 |
JP2020126363A (en) | 2020-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12094151B2 (en) | Image processing system, image processing method, learned model generation method, and data set for learning | |
US11414837B2 (en) | Image processing system, display device, image processing method, method for generating trained model, and dataset for learning | |
AU2022209235B2 (en) | Display control device and display control method | |
US9990543B2 (en) | Vehicle exterior moving object detection system | |
CN110805093B (en) | Container angle sensing with feedback loop control using vision sensors | |
US9133600B2 (en) | Method for selecting an attack pose for a working machine having a bucket | |
CN112513563B (en) | Work machine transported object specifying device, work machine transported object specifying method, completion model production method, and learning dataset | |
JP6867132B2 (en) | Work machine detection processing device and work machine detection processing method | |
CN104302848A (en) | Overhead view system for shovel | |
AU2019292458B2 (en) | Display control system, display control device, and display control method | |
JP7023813B2 (en) | Work machine | |
AU2019292457B2 (en) | Display control device, display control system, and display control method | |
JP7071203B2 (en) | Work machine | |
CN114521228A (en) | Container measuring system | |
JP2024052764A (en) | Display control device and display method | |
US20230267895A1 (en) | Display control device and display control method | |
CN114821496A (en) | Visual covering for providing depth perception | |
Borthwick | Mining haul truck pose estimation and load profiling using stereo vision | |
JP7390991B2 (en) | Work machines and construction support systems |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: KOMATSU LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KAWAMOTO, SHUN;HAMADA, SHINTARO;KAJIHARA, YOSUKE;REEL/FRAME:057029/0299 Effective date: 20210617 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |