US20210272315A1 - Transport object specifying device of work machine, work machine, transport object specifying method of work machine, method for producing complementary model, and dataset for learning - Google Patents
Transport object specifying device of work machine, work machine, transport object specifying method of work machine, method for producing complementary model, and dataset for learning Download PDFInfo
- Publication number
- US20210272315A1 US20210272315A1 US17/260,069 US201917260069A US2021272315A1 US 20210272315 A1 US20210272315 A1 US 20210272315A1 US 201917260069 A US201917260069 A US 201917260069A US 2021272315 A1 US2021272315 A1 US 2021272315A1
- Authority
- US
- United States
- Prior art keywords
- transport object
- drop target
- dump body
- specifying
- dimensional position
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000000295 complement effect Effects 0.000 title claims description 38
- 238000000034 method Methods 0.000 title claims description 23
- 238000004519 manufacturing process Methods 0.000 title claims description 4
- 238000009826 distribution Methods 0.000 claims description 50
- 238000012549 training Methods 0.000 claims description 25
- 239000000284 extract Substances 0.000 claims description 6
- 238000003384 imaging method Methods 0.000 claims description 6
- 230000032258 transport Effects 0.000 description 88
- 238000003860 storage Methods 0.000 description 30
- 210000002569 neuron Anatomy 0.000 description 19
- 238000013528 artificial neural network Methods 0.000 description 14
- 238000010586 diagram Methods 0.000 description 12
- 238000012545 processing Methods 0.000 description 10
- 238000004364 calculation method Methods 0.000 description 4
- 238000005259 measurement Methods 0.000 description 4
- 238000004088 simulation Methods 0.000 description 4
- 238000010276 construction Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000002787 reinforcement Effects 0.000 description 3
- 239000004927 clay Substances 0.000 description 2
- 239000013256 coordination polymer Substances 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000005484 gravity Effects 0.000 description 2
- 239000011435 rock Substances 0.000 description 2
- 239000004576 sand Substances 0.000 description 2
- 239000002023 wood Substances 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 210000004027 cell Anatomy 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- -1 gravel Substances 0.000 description 1
- 239000010720 hydraulic oil Substances 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000000704 physical effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000013526 transfer learning Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- E—FIXED CONSTRUCTIONS
- E02—HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
- E02F—DREDGING; SOIL-SHIFTING
- E02F3/00—Dredgers; Soil-shifting machines
- E02F3/04—Dredgers; Soil-shifting machines mechanically-driven
- E02F3/28—Dredgers; Soil-shifting machines mechanically-driven with digging tools mounted on a dipper- or bucket-arm, i.e. there is either one arm or a pair of arms, e.g. dippers, buckets
- E02F3/36—Component parts
- E02F3/42—Drives for dippers, buckets, dipper-arms or bucket-arms
- E02F3/43—Control of dipper or bucket position; Control of sequence of drive operations
- E02F3/435—Control of dipper or bucket position; Control of sequence of drive operations for dipper-arms, backhoes or the like
- E02F3/439—Automatic repositioning of the implement, e.g. automatic dumping, auto-return
-
- E—FIXED CONSTRUCTIONS
- E02—HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
- E02F—DREDGING; SOIL-SHIFTING
- E02F9/00—Component parts of dredgers or soil-shifting machines, not restricted to one of the kinds covered by groups E02F3/00 - E02F7/00
- E02F9/26—Indicating devices
- E02F9/261—Surveying the work-site to be treated
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/002—Measuring arrangements characterised by the use of optical techniques for measuring two or more coordinates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
Definitions
- the present invention relates to a transport object specifying device of a work machine, a work machine, a transport object specifying method of a work machine, a method for producing a complementary model, and a dataset for learning.
- Japanese Unexamined Patent Application, First Publication No. 2001-71809 discloses a technique of calculating a position of the center of gravity of a transport object based on the output of a weighting sensor provided on a transport vehicle and displaying a loaded state of the transport object.
- the position of the center of gravity of a drop target such as the transport vehicle can be determined, but a three-dimensional position of the transport object in the drop target cannot be specified.
- An object of the present invention is to provide a transport object specifying device of a work machine, a work machine, a transport object specifying method of a work machine, a method for producing a complementary model, and a dataset for learning capable of specifying a three-dimensional position of a transport object in a drop target.
- a transport object specifying device of a work machine includes an image acquisition unit that acquires a captured image showing a drop target of the work machine in which a transport object is dropped, a drop target specifying unit that specifies a three-dimensional position of at least part of the drop target based on the captured image, a three-dimensional data generation unit that generates depth data which is three-dimensional data representing a depth of the captured image, based on the captured image, and a surface specifying unit that specifies a three-dimensional position of a surface of the transport object in the drop target by removing, from the depth data, a part corresponding to the drop target based on the three-dimensional position of the at least part of the drop target.
- the transport object specifying device can specify the distribution of the transport object in the drop target.
- FIG. 1 is a diagram showing a configuration of a loading place according to one embodiment.
- FIG. 2 is an external view of a hydraulic excavator according to one embodiment.
- FIG. 3 is a schematic block diagram showing a configuration of a control device according to a first embodiment.
- FIG. 4 is a diagram showing an example of a configuration of a neural network.
- FIG. 5 is an example of guidance information.
- FIG. 6 is a flowchart showing a display method of the guidance information by the control device according to the first embodiment.
- FIG. 7 is a flowchart showing a learning method of a feature point specifying model according to the first embodiment.
- FIG. 8 is a flowchart showing a learning method of a complementary model according to the first embodiment.
- FIG. 9 is a schematic block diagram showing a configuration of a control device according to a second embodiment.
- FIG. 10 is a flowchart showing a display method of guidance information by the control device according to the second embodiment.
- FIG. 11A is a diagram showing a first example of a method for calculating an amount of a transport object in a dump body.
- FIG. 11B is a diagram showing a second example of the method for calculating the amount of the transport object in the dump body.
- FIG. 1 is a diagram showing a configuration of a loading place according to one embodiment.
- a hydraulic excavator 100 which is a loading machine and a dump truck 200 which is a transport vehicle are provided.
- the hydraulic excavator 100 scoops a transport object L such as earth from the construction site and loads the transport object in the dump truck 200 .
- the dump truck 200 transports the transport object L loaded by the hydraulic excavator 100 to a predetermined earth removable place.
- the dump truck 200 includes a dump body 210 which is a container for accommodating the transport object L.
- the dump body 210 is an example of a drop target in which the transport object L is dropped.
- FIG. 2 is an external view of a hydraulic excavator according to one embodiment.
- the hydraulic excavator 100 includes work equipment 110 that is hydraulically operated, a swing body 120 that supports the work equipment 110 , and a travel body 130 that supports the swing body 120 .
- the swing body 120 is provided with a cab 121 in which an operator rides.
- the cab 121 is provided in a front portion of the swing body 120 and is positioned on a left-side (+Y side) of the work equipment 110 .
- the hydraulic excavator 100 includes a stereo camera 122 , an operation device 123 , a control device 124 , and a display device 125 .
- the stereo camera 122 is provided in an upper portion of the cab 121 .
- the stereo camera 122 is installed in an upper (+Z direction) and front (+X direction) portion of the cab 121 .
- the stereo camera 122 captures an image in front (+X direction) of the cab 121 through a windshield on a front surface of the cab 121 .
- the stereo camera 122 includes at least one pair of cameras.
- the operation device 123 is provided inside the cab 121 .
- the operation device 123 is operated by the operator to supply hydraulic oil to an actuator of the work equipment 110 .
- the control device 124 acquires information from the stereo camera 122 to generate guidance information indicating a distribution of the transport object in the dump body 210 of the dump truck 200 .
- the control device 124 is an example of a transport object specifying device.
- the display device 125 displays the guidance information generated by the control device 124 .
- the hydraulic excavator 100 may not necessarily include the stereo camera 122 and the display device 125 .
- the stereo camera 122 includes a right-side camera 1221 and a left-side camera 1222 .
- Examples of each camera include a camera using a charge coupled device (CCD) sensor and a complementary metal oxide semiconductor (CMOS) sensor.
- CCD charge coupled device
- CMOS complementary metal oxide semiconductor
- the right-side camera 1221 and the left-side camera 1222 are installed at an interval in a left-right direction (Y-axis direction) such that optical axes of the cameras 1221 and 1222 are substantially parallel to a floor surface of the cab 121 .
- the stereo camera 122 is an example of an imaging device.
- the control device 124 can calculate a distance between the stereo camera 122 and a captured target by using an image captured by the right-side camera 1221 and an image captured by the left-side camera 1222 .
- the image captured by the right-side camera 1221 is also referred to as a right-eye image.
- the image captured by the left-side camera 1222 is also referred to as a left-eye image.
- a combination of the images captured by respective cameras of the stereo camera 122 is also referred to as a stereo image.
- the stereo camera 122 may be configured of three or more cameras.
- FIG. 3 is a schematic block diagram showing a configuration of the control device according to the first embodiment.
- the control device 124 includes a processor 91 , a main memory 92 , a storage 93 , and an interface 94 .
- the storage 93 stores a program for controlling the work equipment 110 .
- Examples of the storage 93 include a hard disk drive (HDD) and a non-volatile memory.
- the storage 93 may be an internal medium directly connected to a bus of the control device 124 , or may be an external medium connected to the control device 124 through the interface 94 or a communication line.
- the storage 93 is an example of a storage unit.
- the processor 91 reads the program from the storage 93 , expands the program in the main memory 92 , and executes processing according to the program.
- the processor 91 secures a storage area in the main memory 92 according to the program.
- the interface 94 is connected to the stereo camera 122 , the display device 125 , and other peripheral devices, and transmits and receives signals.
- the main memory 92 is an example of the storage unit.
- the processor 91 includes a data acquisition unit 1701 , a feature point specifying unit 1702 , a three-dimensional data generation unit 1703 , a dump body specifying unit 1704 , a surface specifying unit 1705 , a distribution specifying unit 1706 , a distribution estimation unit 1707 , a guidance information generation unit 1708 , and a display control unit 1709 .
- the storage 93 stores a camera parameter CP, a feature point specifying model M 1 , a complementary model M 2 , and a dump body model VD.
- the camera parameter CP is information indicating a position relationship between the swing body 120 and the right-side camera 1221 and a position relationship between the swing body 120 and the left-side camera 1222 .
- the dump body model VD is a three-dimensional model representing a shape of the dump body 210 .
- three-dimensional data representing a shape of the dump truck 200 may be used instead of the dump body model VD.
- the dump body model VD is an example of a target model.
- the program may be for realizing part of functions to be exerted by the control device 124 .
- the program may exert a function by a combination with another program already stored in the storage 93 or a combination with another program installed in another device.
- the control device 124 may include a custom large scale integrated circuit (LSI) such as a programmable logic device (PLD) in addition to or instead of the above configuration.
- LSI large scale integrated circuit
- PLD programmable logic device
- Examples of the PLD include a programmable array logic (PAL), a generic array logic (GAL), a complex programmable logic device (CPLD), and a field programmable gate array (FPGA).
- PAL programmable array logic
- GAL generic array logic
- CPLD complex programmable logic device
- FPGA field programmable gate array
- the data acquisition unit 1701 acquires the stereo image from the stereo camera 122 through the interface 94 .
- the data acquisition unit 1701 is an example of an image acquisition unit.
- the data acquisition unit 1701 may acquire a stereo image from a stereo camera provided in another work machine, a stereo camera installed at the construction site, or the like.
- the feature point specifying unit 1702 inputs the right-eye image of the stereo image acquired by the data acquisition unit 1701 to the feature point specifying model M 1 stored in the storage 93 to specify positions of a plurality of feature points of the dump body 210 shown in the right-eye image.
- Examples of the feature point of the dump body 210 include upper and lower ends of a front panel of the dump body 210 , an intersection of a guard frame of the front panel and a side gate, and upper and lower ends of a fixed post of a tailgate. That is, the feature point is an example of a predetermined position of the drop target.
- the feature point specifying model M 1 includes a neural network 140 shown in FIG. 4 .
- FIG. 4 is a diagram showing an example of a configuration of the neural network.
- the feature point specifying model M 1 is realized by, for example, a trained model of deep neural network (DNN).
- the trained model is configured of a combination of a training model and a trained parameter.
- the neural network 140 includes an input layer 141 , one or more intermediate layers 142 (hidden layers), and an output layer 143 .
- Each of the layers 141 , 142 , and 143 includes one or more neurons.
- the number of neurons in the intermediate layer 142 can be set as appropriate.
- the output layer 143 can be set as appropriate according to the number of feature points.
- Neurons in the layers adjacent to each other are connected to each other, and a weight (connection load) is set for each connection.
- the number of connected neurons may be set as appropriate.
- a threshold value is set for each neuron, and an output value of each neuron is determined by whether or not a sum of products of an input value and the weight for each neuron exceeds the threshold value.
- An image showing the dump body 210 of the dump truck 200 is input to the input layer 141 .
- an output value indicating a probability of the pixel being the feature point is output to the output layer 143 .
- the feature point specifying model M 1 is a trained model which is trained, when an image showing the dump body 210 is input, to output the positions of the feature points of the dump body 210 in the image.
- the feature point specifying model M 1 is trained by using, for example, a dataset for learning with an image showing the dump body 210 of the dump truck 200 as training data and with an image obtained by plotting the positions of the feature points of the dump body 210 as teaching data.
- the teaching data is an image in which a pixel related to the plot has a value indicating that the probability of the pixel being the feature point is 1, and other pixel has a value indicating that the probability of the pixel being the feature point is 0.
- the teaching data may be information of which a pixel related to the plot has a value indicating that the probability of the pixel being the feature point is 1, and other pixel has a value indicating that the probability of the pixel being the feature point is 0, and may not be an image.
- “training data” refers to data input to the input layer during training of the training model.
- “teaching data” is data which is a correct answer for comparison with the value of the output layer of the neural network 140 .
- “dataset for learning” refers to a combination of the training data and the teaching data.
- the trained parameters of the feature point specifying model M 1 obtained by training are stored in the storage 93 .
- the trained parameters include, for example, the number of layers of the neural network 140 , the number of neurons in each layer, the connection relationship between the neurons, the weight of each connection between the neurons, and the threshold value of each neuron.
- the same or similar DNN configuration as a DNN configuration used for detecting a facial organ or a DNN configuration used for estimating a posture of a person can be used as the configuration of the neural network 140 of the feature point specifying model M 1 .
- the feature point specifying model M 1 is an example of a position specifying model.
- the feature point specifying model M 1 according to another embodiment may be trained by unsupervised learning or reinforcement learning.
- the three-dimensional data generation unit 1703 generates a three-dimensional map representing a depth in an imaging range of the stereo camera 122 by stereo measurement using the stereo image and the camera parameters stored in the storage 93 . Specifically, the three-dimensional data generation unit 1703 generates point group data indicating a three-dimensional position by the stereo measurement of the stereo image.
- the point group data is an example of depth data.
- the three-dimensional data generation unit 1703 may generate an elevation map generated from the point group data as three-dimensional data instead of the point group data.
- the dump body specifying unit 1704 specifies a three-dimensional position of the dump body 210 based on the positions of the feature points specified by the feature point specifying unit 1702 , the point group data specified by the three-dimensional data generation unit 1703 , and the dump body model VD. Specifically, the dump body specifying unit 1704 specifies three-dimensional positions of the feature points based on the positions of the feature points specified by the feature point specifying unit 1702 and the point group data specified by the three-dimensional data generation unit 1703 . Next, the dump body specifying unit 1704 fits the dump body model VD to the three-dimensional positions of the feature points to specify the three-dimensional position of the dump body 210 . In another embodiment, the dump body specifying unit 1704 may specify the three-dimensional position of the dump body 210 based on the elevation map.
- the surface specifying unit 1705 specifies a three-dimensional position of a surface of the transport object L on the dump body 210 based on the point group data generated by the three-dimensional data generation unit 1703 and the three-dimensional position of the dump body 210 specified by the dump body specifying unit 1704 . Specifically, the surface specifying unit 1705 cuts out a part above a bottom surface of the dump body 210 from the point group data generated by the three-dimensional data generation unit 1703 to specify the three-dimensional position of the surface of the transport object L on the dump body 210 .
- the distribution specifying unit 1706 generates a dump body map indicating a distribution of an amount of the transport object L on the dump body 210 based on the three-dimensional position of the bottom surface of the dump body 210 specified by the dump body specifying unit 1704 and the three-dimensional position of the surface of the transport object L specified by the surface specifying unit 1705 .
- the dump body map is an example of distribution information.
- the dump body map is, for example, an elevation map of the transport object L with reference to the bottom surface of the dump body 210 .
- the distribution estimation unit 1707 generates a dump body map in which a value is complemented for a part of the dump body map that does not have a value of height data. That is, the distribution estimation unit 1707 estimates a three-dimensional position of a shielded part of the dump body map that is shielded by an obstacle to update the dump body map. Examples of the obstacle include the work equipment 110 , the tailgate of the dump body 210 , and the transport object L.
- the distribution estimation unit 1707 inputs the dump body map into the complementary model M 2 stored in the storage 93 to generate a dump body map in which the height data is complemented.
- the complementary model M 2 is realized by, for example, a trained model of DNN including the neural network 140 shown in FIG. 4 .
- the complementary model M 2 is a trained model which is trained, when a dump body map including a grid without the height data is input, to output a dump body map in which all grids have the height data.
- the complementary model M 2 is trained with a combination of a complete dump body map in which all grids have the height data, which is generated by simulation or the like, and an incomplete dump body map in which part of the height data is removed from the complete dump body map, as a dataset for learning.
- the complementary model M 2 according to another embodiment may be trained by unsupervised learning or reinforcement learning.
- the guidance information generation unit 1708 generates the guidance information from the dump body map generated by the distribution estimation unit 1707 .
- FIG. 5 is an example of the guidance information.
- the guidance information generation unit 1708 generates the guidance information for displaying a two-dimensional heat map indicating a distribution of the height from the bottom surface of the dump body 210 to the surface of the transport object L.
- Granularity of vertical and horizontal divisions in the heat map shown in FIG. 5 is an example and is not limited thereto in another embodiment.
- the heat map according to another embodiment may represent, for example, a ratio of a height of the transport object L to a height related to an upper limit of the loading of the dump body 210 .
- the display control unit 1709 outputs a display signal for displaying the guidance information to the display device 125 .
- the learning unit 1801 performs learning processing of the feature point specifying model M 1 and the complementary model M 2 .
- the learning unit 1801 may be provided in a device separate from the control device 124 . In this case, the trained model which has been trained in the separate device will be recorded in the storage 93 .
- FIG. 6 is a flowchart showing a display method of the guidance information by the control device according to the first embodiment.
- the data acquisition unit 1701 acquires the stereo image from the stereo camera 122 (step S 1 ).
- the feature point specifying unit 1702 inputs the right-eye image of the stereo image acquired by the data acquisition unit 1701 to the feature point specifying model M 1 stored in the storage 93 to specify the positions of the plurality of feature points of the dump body 210 shown in the right-eye image (step S 2 ).
- the feature point of the dump body 210 include the upper and lower ends of the front panel of the dump body 210 , the intersection of the guard frame of the front panel and the side gate, and the upper and lower ends of the fixed post of the tailgate.
- the feature point specifying unit 1702 may input the left-eye image to the feature point specifying model M 1 to specify the positions of the plurality of feature points.
- the three-dimensional data generation unit 1703 generates the point group data of the entire imaging range of the stereo camera 122 by the stereo measurement using the stereo image acquired in step S 1 and the camera parameters stored in the storage 93 (step S 3 ).
- the dump body specifying unit 1704 specifies the three-dimensional positions of the feature points based on the positions of the feature points specified in step S 2 and the point group data generated in step S 3 (step S 4 ). For example, the dump body specifying unit 1704 specifies, using the point group data, a three-dimensional point corresponding to the pixel showing the feature point in the right-eye image to specify the three-dimensional position of the feature point. The dump body specifying unit 1704 fits the dump body model VD stored in the storage 93 to the specified positions of the feature points to specify the three-dimensional position of the dump body 210 (step S 5 ).
- the dump body specifying unit 1704 may convert a coordinate system of the point group data into a dump body coordinate system having a corner of the dump body 210 as the origin, based on the three-dimensional position of the dump body 210 .
- the dump body coordinate system can be represented as, for example, a coordinate system composed of an X-axis extending in a width direction of the front panel, a Y-axis extending in a width direction of the side gate, and a Z-axis extending in a height direction of the front panel, with a lower left end of the front panel as the origin.
- the dump body specifying unit 1704 is an example of a drop target specifying unit.
- the surface specifying unit 1705 extracts, from the point group data generated in step S 3 , a plurality of three-dimensional points in a prismatic area, which is surrounded by the front panel, the side gate, and the tailgate of the dump body 210 specified in step S 5 and extends in the height direction of the front panel, to remove three-dimensional points corresponding to the background from the point group data (step S 6 ).
- the front panel, the side gate, and the tailgate form a wall portion of the dump body 210 .
- the surface specifying unit 1705 sets threshold values determined based on a known size of the dump body 210 on the X-axis, the Y-axis, and the Z-axis to extract three-dimensional points in an area defined from the thresholds.
- the height of the prismatic area may be equal to the height of the front panel or may be higher than the height of the front panel by a predetermined length.
- the transport object L can be accurately extracted even in a case where the transport object L is loaded higher than the height of the dump body 210 by making a height of the prismatic area higher than that of the front panel.
- the prismatic area may be an area narrowed inward by a predetermined distance from the area surrounded by the front panel, the side gate, and the tailgate.
- the dump body model VD is a simple 3 D model in which thicknesses of the front panel, the side gate, the tailgate, and the bottom surface are not accurate, an error in the point group data can be reduced.
- the surface specifying unit 1705 removes three-dimensional points corresponding to the position of the dump body model VD from the plurality of three-dimensional points extracted in step S 6 to specify the three-dimensional position of the surface of the transport object L loaded on the dump body 210 (step S 7 ).
- the distribution specifying unit 1706 generates the dump body map which is an elevation map representing the height in the height direction of the front panel with the bottom surface of the dump body 210 as a reference height, based on the plurality of three-dimensional points extracted in step S 6 and the bottom surface of the dump body 210 (step S 8 ).
- the dump body map may include a grid without the height data.
- the distribution specifying unit 1706 can generate the dump body map by obtaining an elevation map with an XY plane as the reference height and with the Z-axis direction as the height direction.
- the distribution estimation unit 1707 inputs the dump body map generated in step S 7 into the complementary model M 2 stored in the storage 93 to generate the dump body map in which the height data is complemented (step S 8 ).
- the guidance information generation unit 1708 generates the guidance information shown in FIG. 5 based on the dump body map (step S 9 ).
- the display control unit 1709 outputs the display signal for displaying the guidance information to the display device 125 (step S 10 ).
- steps S 2 to S 4 and steps S 7 to S 10 among the processing by the control device 124 shown in FIG. 6 may not be executed.
- the positions of the feature points in the left-eye image may be specified from the positions of the feature points in the right-eye image by the stereo matching to specify the three-dimensional positions of the feature points using triangulation.
- point group data only in the prismatic area which is surrounded by the front panel, the side gate, and the tailgate of the dump body 210 specified in step S 5 and extends in the height direction of the front panel may be generated. In this case, since it is not necessary to generate the point group data of the entire imaging range, the calculation load can be reduced.
- FIG. 7 is a flowchart showing a learning method of the feature point specifying model M 1 according to the first embodiment.
- the data acquisition unit 1701 acquires the training data (step S 101 ).
- the training data in the feature point specifying model M 1 is an image showing the dump body 210 .
- the training data may be acquired from an image captured by the stereo camera 122 .
- the training data may be acquired from an image captured by another work machine.
- An image showing a work machine different from the dump truck, for example, an image showing a dump body of a wheel loader may be used as the training data. It is possible to improve robustness of dump body recognition by using dump bodies of various types of work machines as the training data.
- the learning unit 1801 performs training of the feature point specifying model M 1 .
- the learning unit 1801 performs training of the feature point specifying model M 1 using the combination of the training data acquired in step S 101 and the teaching data which is the image obtained by plotting the positions of the feature points of the dump body, as the dataset for learning (step S 102 ).
- the learning unit 1801 uses the training data as an input to perform calculation processing of the neural network 140 in a forward propagation direction. Accordingly, the learning unit 1801 obtains an output value output from the output layer 143 of the neural network 140 .
- the dataset for learning may be stored in the main memory 92 or the storage 93 .
- the learning unit 1801 calculates an error between the output value from the output layer 143 and the teaching data.
- the output value from the output layer 143 is a value representing the probability of a pixel being the feature point, and the teaching data is the information obtained by plotting the position of the feature point.
- the learning unit 1801 calculates an error of the weight of each connection between the neurons and an error of the threshold value of each neuron by backpropagation from the calculated error of the output value.
- the learning unit 1801 updates the weight of each connection between the neurons and the threshold value of each neuron based on the calculated errors.
- the learning unit 1801 determines whether or not the output value from the feature point specifying model M 1 matches the teaching data (step S 103 ). It may be determined that the output value matches the teaching data when an error between the output value and the teaching data is within a predetermined value. In a case where the output value from the feature point specifying model M 1 does not match the teaching data (step S 103 : NO), the above processing is repeated until the output value from the feature point specifying model M 1 matches the teaching data. As a result, the parameters of the feature point specifying model M 1 are optimized, and the feature point specifying model M 1 can be trained.
- the learning unit 1801 records the feature point specifying model M 1 as a trained model including the parameters optimized by the training in the storage 93 (step S 104 ).
- FIG. 8 is a flowchart showing a learning method of the complementary model according to the first embodiment.
- the data acquisition unit 1701 acquires the complete dump body map in which all grids have the height data as teaching data (step S 111 ).
- the complete dump body map is generated, for example, by simulation or the like.
- the learning unit 1801 randomly removes a part of the height data of the complete dump body map to generate the incomplete dump body map as training data (step S 112 ).
- the learning unit 1801 performs training of the complementary model M 2 .
- the learning unit 1801 performs training of the complementary model M 2 with the combination of the training data generated in step S 112 and the teaching data acquired in step S 111 as the dataset for learning (step S 113 ).
- the learning unit 1801 uses the training data as an input to perform calculation processing of the neural network 140 in a forward propagation direction. Accordingly, the learning unit 1801 obtains an output value output from the output layer 143 of the neural network 140 .
- the dataset for learning may be stored in the main memory 92 or the storage 93 .
- the learning unit 1801 calculates an error between the dump body map output from the output layer 143 and the complete dump body map as the teaching data.
- the learning unit 1801 calculates an error of the weight of each connection between the neurons and an error of threshold value of each neuron by backpropagation from the calculated error of the output value.
- the learning unit 1801 updates the weight of each connection between the neurons and the threshold value of each neuron based on the calculated errors.
- the learning unit 1801 determines whether or not the output value from the complementary model M 2 matches the teaching data (step S 114 ). It may be determined that the output value matches the teaching data when an error between the output value and the teaching data is within a predetermined value. In a case where the output value from the complementary model M 2 does not match the teaching data (step S 114 : NO), the above processing is repeated until the output value from the complementary model M 2 matches the complete dump body map. As a result, the parameters of the complementary model M 2 are optimized, and the complementary model M 2 can be trained.
- the learning unit 1801 records the complementary model M 2 as a trained model including the parameters optimized by the training in the storage 93 (step S 115 ).
- the control device 124 specifies the three-dimensional positions of the surface of the transport object L and the bottom surface of the dump body 210 based on the captured image, and generates the dump body map indicating the distribution of the amount of the transport object L on the dump body 210 based on the three-dimensional positions. Accordingly, the control device 124 can specify the distribution of the transport object L on the dump body 210 . The operator can recognize the drop position of the transport object L for loading the transport object L on the dump body 210 in a well-balanced manner by recognizing the distribution of the transport object L on the dump body 210 .
- the control device 124 estimates the distribution of the amount of the transport object L in the shielded part of the dump body map shielded by an obstacle. Accordingly, the operator can recognize the distribution of the amount of the transport object L even for a part of the dump body 210 that is shielded by the obstacle and cannot be captured by the stereo camera 122 .
- the control device 124 specifies the distribution of the transport object L on the dump body 210 based on a type of the transport object L.
- FIG. 9 is a schematic block diagram showing a configuration of a control device according to the second embodiment.
- the control device 124 further includes a type specifying unit 1710 .
- the storage 93 stores a type specifying model M 3 and a plurality of complementary models M 2 according to the type of the transport object L.
- the type specifying unit 1710 inputs an image of the transport object L to the type specifying model M 3 to specify the type of the transport object L shown in the image.
- Examples of the type of transport object include clay, sand, gravel, rock, and wood.
- the type specifying model M 3 is realized by, for example, a trained model of deep neural network (DNN).
- the type specifying model M 3 is a trained model which is trained, when an image showing the transport object L is input, to output the type of the transport object L.
- a DNN configuration of the type specifying model M 3 for example, the same or similar DNN configuration as a DNN configuration used for image recognition can be used.
- the type specifying model M 3 is trained, for example, using a combination of an image showing the transport object L and a label representing the type of the transport object L as teaching data.
- the type specifying model M 3 is trained using a combination of an image showing the transport object L and label data representing the type of the transport object L as the teaching data.
- the type specifying model M 3 may be trained by transfer learning of a general trained image recognition model.
- the type specifying model M 3 according to another embodiment may be trained by unsupervised learning or reinforcement learning.
- the storage 93 stores the complementary model M 2 for each type of the transport object L.
- the storage 93 stores a complementary model M 2 for clay, a complementary model M 2 for sand, a complementary model M 2 for gravel, a complementary model M 2 for rock, and a complementary model M 2 for wood.
- Each complementary model M 2 is trained, for example, using a combination of a complete dump body map generated by simulation or the like according to the type of the transport object L and an incomplete dump body map obtained by removing part of the height data from the dump body map as teaching data.
- FIG. 10 is a flowchart showing a display method of the guidance information by the control device according to the second embodiment.
- the data acquisition unit 1701 acquires the stereo image from the stereo camera 122 (step S 21 ).
- the feature point specifying unit 1702 inputs the right-eye image of the stereo image acquired by the data acquisition unit 1701 to the feature point specifying model M 1 stored in the storage 93 to specify the positions of the plurality of feature points of the dump body 210 shown in the right-eye image (step S 22 ).
- the three-dimensional data generation unit 1703 generates the point group data of the entire imaging range of the stereo camera 122 by the stereo measurement using the stereo image acquired in step S 21 and the camera parameters stored in the storage 93 (step S 23 ).
- the dump body specifying unit 1704 specifies the three-dimensional positions of the feature points based on the positions of the feature points specified in step S 22 and the point group data generated in step S 23 (step S 24 ).
- the dump body specifying unit 1704 fits the dump body model VD stored in the storage 93 to the specified positions of the feature points to specify the three-dimensional position of the bottom surface of the dump body 210 (step S 25 ).
- the dump body specifying unit 1704 disposes, based on at least three specified positions of the feature points, the dump body model VD created based on the dimensions of the dump truck 200 to be detected in a virtual space.
- the surface specifying unit 1705 extracts, from the point group data generated in step S 23 , a plurality of three-dimensional points in a prismatic area, which is surrounded by the front panel, the side gate, and the tailgate of the dump body 210 specified in step S 25 and extends in the height direction of the front panel, to remove three-dimensional points corresponding to the background from the point group data (step S 26 ).
- the surface specifying unit 1705 removes three-dimensional points corresponding to the position of the dump body model VD from the plurality of three-dimensional points extracted in step S 6 to specify the three-dimensional position of the surface of the transport object L loaded on the dump body 210 (step S 27 ).
- the distribution specifying unit 1706 generates the dump body map which is an elevation map with the bottom surface of the dump body 210 as a reference height, based on the plurality of three-dimensional points extracted in step S 27 and the bottom surface of the dump body 210 (step S 28 ).
- the dump body map may include a grid without the height data.
- the surface specifying unit 1705 specifies an area where the transport object L appears in the right-eye image based on the three-dimensional position of the surface of the transport object L specified in step S 27 (step S 29 ). For example, the surface specifying unit 1705 specifies a plurality of pixels in the right-eye image corresponding to the plurality of three-dimensional points extracted in step S 27 and determines an area composed of the plurality of specified pixels as the area where the transport object L appears.
- the type specifying unit 1710 extracts the area where the transport object L appears from the right-eye image and inputs an image related to the area to the type specifying model M 3 to specify the type of the transport object L (step S 30 ).
- the distribution estimation unit 1707 inputs the dump body map generated in step S 28 to the complementary model M 2 associated with the type specified in step S 30 to generate the dump body map in which the height data is complemented (step S 31 ).
- the guidance information generation unit 1708 generates the guidance information based on the dump body map (step S 32 ).
- the display control unit 1709 outputs the display signal for displaying the guidance information to the display device 125 (step S 33 ).
- the control device 124 estimates the distribution of the amount of the transport object L in the shielded part based on the type of the transport object L. That is, characteristics (for example, the angle of repose) of the transport object L loaded on the dump body 210 differ depending on the type of the transport object L.
- the third embodiment it is possible to more accurately estimate the distribution of the transport object L in the shielded part according to the type of the transport object L.
- control device 124 is mounted on the hydraulic excavator 100
- the present invention is not limited thereto.
- the control device 124 according to another embodiment may be provided in a remote server device.
- the control device 124 may be realized by a plurality of computers. In this case, part of the configuration of the control device 124 may be provided in the remote server device. That is, the control device 124 may be implemented as a transport object specifying system composed of a plurality of devices.
- the drop target according to the above-described embodiment is the dump body 210 of the dump truck 200
- the present invention is not limited thereto.
- the drop target according to another embodiment may be another drop target such as a hopper.
- the captured image according to the above-described embodiment is the stereo image
- the present invention is not limited thereto.
- the calculation may be performed based on one image instead of the stereo image.
- the control device 124 can specify the three-dimensional position of the transport object L by using, for example, a trained model that generates depth information from the one image.
- control device 124 complements the value of the shielded part of the dump body map by using the complementary model M 2
- the present invention is not limited thereto.
- the control device 124 according to another embodiment may estimate a height of the shielded part based on a rate of change or a pattern of change in the height of the transport object L near the shielded part. For example, in a case where the height of the transport object L near the shielded part becomes lower as it approaches the shielded part, the control device 124 may estimate the height of the transport object L in the shielded part to a value lower than the height near the shielded part based on the rate of change in the height.
- the control device 124 may estimate the height of the transport object L in the shielded part by simulation in consideration of a physical property such as the angle of repose of the transport object L.
- the control device 124 according to another embodiment may deterministically estimate the height of the transport object L in the shielded part based on cellular automaton in which each grid of the dump body map is regarded as a cell.
- the control device 124 may not complement the dump body map and may display information related to the dump body map including a part where the height data is missing.
- FIG. 11A is a diagram showing a first example of a method for calculating an amount of a transport object in a dump body.
- FIG. 11B is a diagram showing a second example of the method for calculating the amount of the transport object in the dump body.
- the dump body map according to the above-described embodiment is represented by a height from a bottom surface L 1 of the dump body 210 to an upper limit of the loading on the dump body 210 , but the present invention is not limited thereto.
- the dump body map may represent a height from another reference plane L 3 with respect to the bottom surface to a surface L 2 of the transport object L.
- the reference plane L 3 is a plane parallel to the ground surface and passing through a point of the bottom surface closest to the ground surface. In this case, the operator can easily recognize the amount of the transport object L until the dump body 210 is full, regardless of an inclination of the dump body 210 .
- control device 124 may calculate the dump body map based on an opening surface of the dump body 210 , the surface of the transport object, and a height from the bottom surface to the opening surface of the dump body 210 . That is, the control device 124 may calculate the dump body map by subtracting, from the height from the bottom surface to the opening surface of the dump body 210 , a distance from an upper end surface of the dump body to the surface of the transport object L.
- the dump body map according to another embodiment may be based on the opening surface of the dump body 210 .
- the guidance information generation unit 1708 extracts the feature points from the right-eye image using the feature point specifying model M 1 , the present invention is not limited thereto.
- the guidance information generation unit 1708 may extract the feature points from the left-eye image using the feature point specifying model M 1 .
- the transport object specifying device can specify the distribution of the transport object in the drop target.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Theoretical Computer Science (AREA)
- Mining & Mineral Resources (AREA)
- Mechanical Engineering (AREA)
- Civil Engineering (AREA)
- General Engineering & Computer Science (AREA)
- Structural Engineering (AREA)
- Image Analysis (AREA)
- Component Parts Of Construction Machinery (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
A transport object specifying device of a work machine includes an image acquisition unit, a drop target specifying unit, a three-dimensional data generation unit, and a surface specifying unit. The image acquisition unit acquires a captured image showing a drop target of the work machine in which a transport object is dropped. The drop target specifying unit specifies a three-dimensional position of at least part of the drop target based on the captured image. The three-dimensional data generation unit generates depth data, which is three-dimensional data representing a depth of the captured image, based on the captured image. The surface specifying unit specifies a three-dimensional position of a surface of the transport object in the drop target by removing, from the depth data, a part corresponding to the drop target based on the three-dimensional position of the at least part of the drop target.
Description
- This application is a U.S. National stage application of International Application No. PCT/JP2019/028454, filed on Jul. 19, 2019. This U.S. National stage application claims priority under 35 U.S.C. § 119(a) to Japanese Patent Application No. 2018-163671, filed in Japan on Aug. 31, 2018, the entire contents of which are hereby incorporated herein by reference.
- The present invention relates to a transport object specifying device of a work machine, a work machine, a transport object specifying method of a work machine, a method for producing a complementary model, and a dataset for learning.
- Japanese Unexamined Patent Application, First Publication No. 2001-71809 discloses a technique of calculating a position of the center of gravity of a transport object based on the output of a weighting sensor provided on a transport vehicle and displaying a loaded state of the transport object.
- In the method described in Japanese Unexamined Patent Application, First Publication No. 2001-71809, the position of the center of gravity of a drop target such as the transport vehicle can be determined, but a three-dimensional position of the transport object in the drop target cannot be specified.
- An object of the present invention is to provide a transport object specifying device of a work machine, a work machine, a transport object specifying method of a work machine, a method for producing a complementary model, and a dataset for learning capable of specifying a three-dimensional position of a transport object in a drop target.
- According to one aspect of the present invention, a transport object specifying device of a work machine includes an image acquisition unit that acquires a captured image showing a drop target of the work machine in which a transport object is dropped, a drop target specifying unit that specifies a three-dimensional position of at least part of the drop target based on the captured image, a three-dimensional data generation unit that generates depth data which is three-dimensional data representing a depth of the captured image, based on the captured image, and a surface specifying unit that specifies a three-dimensional position of a surface of the transport object in the drop target by removing, from the depth data, a part corresponding to the drop target based on the three-dimensional position of the at least part of the drop target.
- According to at least one of the above aspects, the transport object specifying device can specify the distribution of the transport object in the drop target.
-
FIG. 1 is a diagram showing a configuration of a loading place according to one embodiment. -
FIG. 2 is an external view of a hydraulic excavator according to one embodiment. -
FIG. 3 is a schematic block diagram showing a configuration of a control device according to a first embodiment. -
FIG. 4 is a diagram showing an example of a configuration of a neural network. -
FIG. 5 is an example of guidance information. -
FIG. 6 is a flowchart showing a display method of the guidance information by the control device according to the first embodiment. -
FIG. 7 is a flowchart showing a learning method of a feature point specifying model according to the first embodiment. -
FIG. 8 is a flowchart showing a learning method of a complementary model according to the first embodiment. -
FIG. 9 is a schematic block diagram showing a configuration of a control device according to a second embodiment. -
FIG. 10 is a flowchart showing a display method of guidance information by the control device according to the second embodiment. -
FIG. 11A is a diagram showing a first example of a method for calculating an amount of a transport object in a dump body. -
FIG. 11B is a diagram showing a second example of the method for calculating the amount of the transport object in the dump body. - Hereinafter, embodiments will be described in detail with reference to drawings.
-
FIG. 1 is a diagram showing a configuration of a loading place according to one embodiment. - At a construction site, a
hydraulic excavator 100 which is a loading machine and adump truck 200 which is a transport vehicle are provided. Thehydraulic excavator 100 scoops a transport object L such as earth from the construction site and loads the transport object in thedump truck 200. Thedump truck 200 transports the transport object L loaded by thehydraulic excavator 100 to a predetermined earth removable place. Thedump truck 200 includes adump body 210 which is a container for accommodating the transport object L. Thedump body 210 is an example of a drop target in which the transport object L is dropped. -
FIG. 2 is an external view of a hydraulic excavator according to one embodiment. - The
hydraulic excavator 100 includeswork equipment 110 that is hydraulically operated, aswing body 120 that supports thework equipment 110, and atravel body 130 that supports theswing body 120. - The
swing body 120 is provided with acab 121 in which an operator rides. Thecab 121 is provided in a front portion of theswing body 120 and is positioned on a left-side (+Y side) of thework equipment 110. - The
hydraulic excavator 100 includes astereo camera 122, anoperation device 123, acontrol device 124, and adisplay device 125. - The
stereo camera 122 is provided in an upper portion of thecab 121. Thestereo camera 122 is installed in an upper (+Z direction) and front (+X direction) portion of thecab 121. Thestereo camera 122 captures an image in front (+X direction) of thecab 121 through a windshield on a front surface of thecab 121. Thestereo camera 122 includes at least one pair of cameras. - The
operation device 123 is provided inside thecab 121. Theoperation device 123 is operated by the operator to supply hydraulic oil to an actuator of thework equipment 110. - The
control device 124 acquires information from thestereo camera 122 to generate guidance information indicating a distribution of the transport object in thedump body 210 of thedump truck 200. Thecontrol device 124 is an example of a transport object specifying device. - The
display device 125 displays the guidance information generated by thecontrol device 124. - The
hydraulic excavator 100 according to another embodiment may not necessarily include thestereo camera 122 and thedisplay device 125. - In the first embodiment, the
stereo camera 122 includes a right-side camera 1221 and a left-side camera 1222. Examples of each camera include a camera using a charge coupled device (CCD) sensor and a complementary metal oxide semiconductor (CMOS) sensor. - The right-
side camera 1221 and the left-side camera 1222 are installed at an interval in a left-right direction (Y-axis direction) such that optical axes of thecameras cab 121. Thestereo camera 122 is an example of an imaging device. Thecontrol device 124 can calculate a distance between thestereo camera 122 and a captured target by using an image captured by the right-side camera 1221 and an image captured by the left-side camera 1222. Hereinafter, the image captured by the right-side camera 1221 is also referred to as a right-eye image. The image captured by the left-side camera 1222 is also referred to as a left-eye image. A combination of the images captured by respective cameras of thestereo camera 122 is also referred to as a stereo image. In another embodiment, thestereo camera 122 may be configured of three or more cameras. -
FIG. 3 is a schematic block diagram showing a configuration of the control device according to the first embodiment. - The
control device 124 includes aprocessor 91, amain memory 92, astorage 93, and aninterface 94. - The
storage 93 stores a program for controlling thework equipment 110. Examples of thestorage 93 include a hard disk drive (HDD) and a non-volatile memory. Thestorage 93 may be an internal medium directly connected to a bus of thecontrol device 124, or may be an external medium connected to thecontrol device 124 through theinterface 94 or a communication line. Thestorage 93 is an example of a storage unit. - The
processor 91 reads the program from thestorage 93, expands the program in themain memory 92, and executes processing according to the program. Theprocessor 91 secures a storage area in themain memory 92 according to the program. Theinterface 94 is connected to thestereo camera 122, thedisplay device 125, and other peripheral devices, and transmits and receives signals. Themain memory 92 is an example of the storage unit. - With the execution of the program, the
processor 91 includes adata acquisition unit 1701, a featurepoint specifying unit 1702, a three-dimensionaldata generation unit 1703, a dumpbody specifying unit 1704, asurface specifying unit 1705, adistribution specifying unit 1706, adistribution estimation unit 1707, a guidanceinformation generation unit 1708, and adisplay control unit 1709. Thestorage 93 stores a camera parameter CP, a feature point specifying model M1, a complementary model M2, and a dump body model VD. The camera parameter CP is information indicating a position relationship between theswing body 120 and the right-side camera 1221 and a position relationship between theswing body 120 and the left-side camera 1222. The dump body model VD is a three-dimensional model representing a shape of thedump body 210. In another embodiment, three-dimensional data representing a shape of thedump truck 200 may be used instead of the dump body model VD. The dump body model VD is an example of a target model. - The program may be for realizing part of functions to be exerted by the
control device 124. For example, the program may exert a function by a combination with another program already stored in thestorage 93 or a combination with another program installed in another device. In another embodiment, thecontrol device 124 may include a custom large scale integrated circuit (LSI) such as a programmable logic device (PLD) in addition to or instead of the above configuration. Examples of the PLD include a programmable array logic (PAL), a generic array logic (GAL), a complex programmable logic device (CPLD), and a field programmable gate array (FPGA). In this case, some or all of the functions realized by the processor may be realized by the integrated circuit. - The
data acquisition unit 1701 acquires the stereo image from thestereo camera 122 through theinterface 94. Thedata acquisition unit 1701 is an example of an image acquisition unit. In another embodiment, in a case where thehydraulic excavator 100 does not include thestereo camera 122, thedata acquisition unit 1701 may acquire a stereo image from a stereo camera provided in another work machine, a stereo camera installed at the construction site, or the like. - The feature
point specifying unit 1702 inputs the right-eye image of the stereo image acquired by thedata acquisition unit 1701 to the feature point specifying model M1 stored in thestorage 93 to specify positions of a plurality of feature points of thedump body 210 shown in the right-eye image. Examples of the feature point of thedump body 210 include upper and lower ends of a front panel of thedump body 210, an intersection of a guard frame of the front panel and a side gate, and upper and lower ends of a fixed post of a tailgate. That is, the feature point is an example of a predetermined position of the drop target. - The feature point specifying model M1 includes a neural network 140 shown in
FIG. 4 .FIG. 4 is a diagram showing an example of a configuration of the neural network. The feature point specifying model M1 is realized by, for example, a trained model of deep neural network (DNN). The trained model is configured of a combination of a training model and a trained parameter. - As shown in
FIG. 4 , the neural network 140 includes an input layer 141, one or more intermediate layers 142 (hidden layers), and an output layer 143. Each of the layers 141, 142, and 143 includes one or more neurons. The number of neurons in the intermediate layer 142 can be set as appropriate. The output layer 143 can be set as appropriate according to the number of feature points. - Neurons in the layers adjacent to each other are connected to each other, and a weight (connection load) is set for each connection. The number of connected neurons may be set as appropriate. A threshold value is set for each neuron, and an output value of each neuron is determined by whether or not a sum of products of an input value and the weight for each neuron exceeds the threshold value.
- An image showing the
dump body 210 of thedump truck 200 is input to the input layer 141. For each pixel of the image, an output value indicating a probability of the pixel being the feature point is output to the output layer 143. That is, the feature point specifying model M1 is a trained model which is trained, when an image showing thedump body 210 is input, to output the positions of the feature points of thedump body 210 in the image. The feature point specifying model M1 is trained by using, for example, a dataset for learning with an image showing thedump body 210 of thedump truck 200 as training data and with an image obtained by plotting the positions of the feature points of thedump body 210 as teaching data. The teaching data is an image in which a pixel related to the plot has a value indicating that the probability of the pixel being the feature point is 1, and other pixel has a value indicating that the probability of the pixel being the feature point is 0. The teaching data may be information of which a pixel related to the plot has a value indicating that the probability of the pixel being the feature point is 1, and other pixel has a value indicating that the probability of the pixel being the feature point is 0, and may not be an image. In the present embodiment, “training data” refers to data input to the input layer during training of the training model. In the present embodiment, “teaching data” is data which is a correct answer for comparison with the value of the output layer of the neural network 140. In the present embodiment, “dataset for learning” refers to a combination of the training data and the teaching data. The trained parameters of the feature point specifying model M1 obtained by training are stored in thestorage 93. The trained parameters include, for example, the number of layers of the neural network 140, the number of neurons in each layer, the connection relationship between the neurons, the weight of each connection between the neurons, and the threshold value of each neuron. For example, the same or similar DNN configuration as a DNN configuration used for detecting a facial organ or a DNN configuration used for estimating a posture of a person can be used as the configuration of the neural network 140 of the feature point specifying model M1. The feature point specifying model M1 is an example of a position specifying model. The feature point specifying model M1 according to another embodiment may be trained by unsupervised learning or reinforcement learning. - The three-dimensional
data generation unit 1703 generates a three-dimensional map representing a depth in an imaging range of thestereo camera 122 by stereo measurement using the stereo image and the camera parameters stored in thestorage 93. Specifically, the three-dimensionaldata generation unit 1703 generates point group data indicating a three-dimensional position by the stereo measurement of the stereo image. The point group data is an example of depth data. In another embodiment, the three-dimensionaldata generation unit 1703 may generate an elevation map generated from the point group data as three-dimensional data instead of the point group data. - The dump
body specifying unit 1704 specifies a three-dimensional position of thedump body 210 based on the positions of the feature points specified by the featurepoint specifying unit 1702, the point group data specified by the three-dimensionaldata generation unit 1703, and the dump body model VD. Specifically, the dumpbody specifying unit 1704 specifies three-dimensional positions of the feature points based on the positions of the feature points specified by the featurepoint specifying unit 1702 and the point group data specified by the three-dimensionaldata generation unit 1703. Next, the dumpbody specifying unit 1704 fits the dump body model VD to the three-dimensional positions of the feature points to specify the three-dimensional position of thedump body 210. In another embodiment, the dumpbody specifying unit 1704 may specify the three-dimensional position of thedump body 210 based on the elevation map. - The
surface specifying unit 1705 specifies a three-dimensional position of a surface of the transport object L on thedump body 210 based on the point group data generated by the three-dimensionaldata generation unit 1703 and the three-dimensional position of thedump body 210 specified by the dumpbody specifying unit 1704. Specifically, thesurface specifying unit 1705 cuts out a part above a bottom surface of thedump body 210 from the point group data generated by the three-dimensionaldata generation unit 1703 to specify the three-dimensional position of the surface of the transport object L on thedump body 210. - The
distribution specifying unit 1706 generates a dump body map indicating a distribution of an amount of the transport object L on thedump body 210 based on the three-dimensional position of the bottom surface of thedump body 210 specified by the dumpbody specifying unit 1704 and the three-dimensional position of the surface of the transport object L specified by thesurface specifying unit 1705. The dump body map is an example of distribution information. The dump body map is, for example, an elevation map of the transport object L with reference to the bottom surface of thedump body 210. - The
distribution estimation unit 1707 generates a dump body map in which a value is complemented for a part of the dump body map that does not have a value of height data. That is, thedistribution estimation unit 1707 estimates a three-dimensional position of a shielded part of the dump body map that is shielded by an obstacle to update the dump body map. Examples of the obstacle include thework equipment 110, the tailgate of thedump body 210, and the transport object L. - Specifically, the
distribution estimation unit 1707 inputs the dump body map into the complementary model M2 stored in thestorage 93 to generate a dump body map in which the height data is complemented. The complementary model M2 is realized by, for example, a trained model of DNN including the neural network 140 shown inFIG. 4 . The complementary model M2 is a trained model which is trained, when a dump body map including a grid without the height data is input, to output a dump body map in which all grids have the height data. For example, the complementary model M2 is trained with a combination of a complete dump body map in which all grids have the height data, which is generated by simulation or the like, and an incomplete dump body map in which part of the height data is removed from the complete dump body map, as a dataset for learning. The complementary model M2 according to another embodiment may be trained by unsupervised learning or reinforcement learning. - The guidance
information generation unit 1708 generates the guidance information from the dump body map generated by thedistribution estimation unit 1707. -
FIG. 5 is an example of the guidance information. As shown inFIG. 5 , for example, the guidanceinformation generation unit 1708 generates the guidance information for displaying a two-dimensional heat map indicating a distribution of the height from the bottom surface of thedump body 210 to the surface of the transport object L. Granularity of vertical and horizontal divisions in the heat map shown inFIG. 5 is an example and is not limited thereto in another embodiment. The heat map according to another embodiment may represent, for example, a ratio of a height of the transport object L to a height related to an upper limit of the loading of thedump body 210. - The
display control unit 1709 outputs a display signal for displaying the guidance information to thedisplay device 125. - The
learning unit 1801 performs learning processing of the feature point specifying model M1 and the complementary model M2. Thelearning unit 1801 may be provided in a device separate from thecontrol device 124. In this case, the trained model which has been trained in the separate device will be recorded in thestorage 93. -
FIG. 6 is a flowchart showing a display method of the guidance information by the control device according to the first embodiment. - First, the
data acquisition unit 1701 acquires the stereo image from the stereo camera 122 (step S1). Next, the featurepoint specifying unit 1702 inputs the right-eye image of the stereo image acquired by thedata acquisition unit 1701 to the feature point specifying model M1 stored in thestorage 93 to specify the positions of the plurality of feature points of thedump body 210 shown in the right-eye image (step S2). Examples of the feature point of thedump body 210 include the upper and lower ends of the front panel of thedump body 210, the intersection of the guard frame of the front panel and the side gate, and the upper and lower ends of the fixed post of the tailgate. In another embodiment, the featurepoint specifying unit 1702 may input the left-eye image to the feature point specifying model M1 to specify the positions of the plurality of feature points. - The three-dimensional
data generation unit 1703 generates the point group data of the entire imaging range of thestereo camera 122 by the stereo measurement using the stereo image acquired in step S1 and the camera parameters stored in the storage 93 (step S3). - The dump
body specifying unit 1704 specifies the three-dimensional positions of the feature points based on the positions of the feature points specified in step S2 and the point group data generated in step S3 (step S4). For example, the dumpbody specifying unit 1704 specifies, using the point group data, a three-dimensional point corresponding to the pixel showing the feature point in the right-eye image to specify the three-dimensional position of the feature point. The dumpbody specifying unit 1704 fits the dump body model VD stored in thestorage 93 to the specified positions of the feature points to specify the three-dimensional position of the dump body 210 (step S5). At this time, the dumpbody specifying unit 1704 may convert a coordinate system of the point group data into a dump body coordinate system having a corner of thedump body 210 as the origin, based on the three-dimensional position of thedump body 210. The dump body coordinate system can be represented as, for example, a coordinate system composed of an X-axis extending in a width direction of the front panel, a Y-axis extending in a width direction of the side gate, and a Z-axis extending in a height direction of the front panel, with a lower left end of the front panel as the origin. The dumpbody specifying unit 1704 is an example of a drop target specifying unit. - The
surface specifying unit 1705 extracts, from the point group data generated in step S3, a plurality of three-dimensional points in a prismatic area, which is surrounded by the front panel, the side gate, and the tailgate of thedump body 210 specified in step S5 and extends in the height direction of the front panel, to remove three-dimensional points corresponding to the background from the point group data (step S6). The front panel, the side gate, and the tailgate form a wall portion of thedump body 210. In a case where the point group data is converted into the dump body coordinate system in step S5, thesurface specifying unit 1705 sets threshold values determined based on a known size of thedump body 210 on the X-axis, the Y-axis, and the Z-axis to extract three-dimensional points in an area defined from the thresholds. The height of the prismatic area may be equal to the height of the front panel or may be higher than the height of the front panel by a predetermined length. The transport object L can be accurately extracted even in a case where the transport object L is loaded higher than the height of thedump body 210 by making a height of the prismatic area higher than that of the front panel. The prismatic area may be an area narrowed inward by a predetermined distance from the area surrounded by the front panel, the side gate, and the tailgate. In this case, even though the dump body model VD is a simple 3D model in which thicknesses of the front panel, the side gate, the tailgate, and the bottom surface are not accurate, an error in the point group data can be reduced. - The
surface specifying unit 1705 removes three-dimensional points corresponding to the position of the dump body model VD from the plurality of three-dimensional points extracted in step S6 to specify the three-dimensional position of the surface of the transport object L loaded on the dump body 210 (step S7). Thedistribution specifying unit 1706 generates the dump body map which is an elevation map representing the height in the height direction of the front panel with the bottom surface of thedump body 210 as a reference height, based on the plurality of three-dimensional points extracted in step S6 and the bottom surface of the dump body 210 (step S8). The dump body map may include a grid without the height data. In a case where the point group data is converted into the dump body coordinate system in step S5, thedistribution specifying unit 1706 can generate the dump body map by obtaining an elevation map with an XY plane as the reference height and with the Z-axis direction as the height direction. - The
distribution estimation unit 1707 inputs the dump body map generated in step S7 into the complementary model M2 stored in thestorage 93 to generate the dump body map in which the height data is complemented (step S8). The guidanceinformation generation unit 1708 generates the guidance information shown inFIG. 5 based on the dump body map (step S9). Thedisplay control unit 1709 outputs the display signal for displaying the guidance information to the display device 125 (step S10). - Depending on the embodiment, the processing of steps S2 to S4 and steps S7 to S10 among the processing by the
control device 124 shown inFIG. 6 may not be executed. - Instead of the processing of steps S3 and S4 among the processing by the
control device 124 shown inFIG. 6 , the positions of the feature points in the left-eye image may be specified from the positions of the feature points in the right-eye image by the stereo matching to specify the three-dimensional positions of the feature points using triangulation. Instead of the process of step S6, point group data only in the prismatic area which is surrounded by the front panel, the side gate, and the tailgate of thedump body 210 specified in step S5 and extends in the height direction of the front panel may be generated. In this case, since it is not necessary to generate the point group data of the entire imaging range, the calculation load can be reduced. -
FIG. 7 is a flowchart showing a learning method of the feature point specifying model M1 according to the first embodiment. Thedata acquisition unit 1701 acquires the training data (step S101). For example, the training data in the feature point specifying model M1 is an image showing thedump body 210. The training data may be acquired from an image captured by thestereo camera 122. The training data may be acquired from an image captured by another work machine. An image showing a work machine different from the dump truck, for example, an image showing a dump body of a wheel loader may be used as the training data. It is possible to improve robustness of dump body recognition by using dump bodies of various types of work machines as the training data. - Next, the
learning unit 1801 performs training of the feature point specifying model M1. Thelearning unit 1801 performs training of the feature point specifying model M1 using the combination of the training data acquired in step S101 and the teaching data which is the image obtained by plotting the positions of the feature points of the dump body, as the dataset for learning (step S102). For example, thelearning unit 1801 uses the training data as an input to perform calculation processing of the neural network 140 in a forward propagation direction. Accordingly, thelearning unit 1801 obtains an output value output from the output layer 143 of the neural network 140. The dataset for learning may be stored in themain memory 92 or thestorage 93. Next, thelearning unit 1801 calculates an error between the output value from the output layer 143 and the teaching data. The output value from the output layer 143 is a value representing the probability of a pixel being the feature point, and the teaching data is the information obtained by plotting the position of the feature point. Thelearning unit 1801 calculates an error of the weight of each connection between the neurons and an error of the threshold value of each neuron by backpropagation from the calculated error of the output value. Thelearning unit 1801 updates the weight of each connection between the neurons and the threshold value of each neuron based on the calculated errors. - The
learning unit 1801 determines whether or not the output value from the feature point specifying model M1 matches the teaching data (step S103). It may be determined that the output value matches the teaching data when an error between the output value and the teaching data is within a predetermined value. In a case where the output value from the feature point specifying model M1 does not match the teaching data (step S103: NO), the above processing is repeated until the output value from the feature point specifying model M1 matches the teaching data. As a result, the parameters of the feature point specifying model M1 are optimized, and the feature point specifying model M1 can be trained. - In a case where the output value from the feature point specifying model M1 matches a value corresponding to the feature point (step S103: YES), the
learning unit 1801 records the feature point specifying model M1 as a trained model including the parameters optimized by the training in the storage 93 (step S104). -
FIG. 8 is a flowchart showing a learning method of the complementary model according to the first embodiment. Thedata acquisition unit 1701 acquires the complete dump body map in which all grids have the height data as teaching data (step S111). The complete dump body map is generated, for example, by simulation or the like. Thelearning unit 1801 randomly removes a part of the height data of the complete dump body map to generate the incomplete dump body map as training data (step S112). - Next, the
learning unit 1801 performs training of the complementary model M2. Thelearning unit 1801 performs training of the complementary model M2 with the combination of the training data generated in step S112 and the teaching data acquired in step S111 as the dataset for learning (step S113). For example, thelearning unit 1801 uses the training data as an input to perform calculation processing of the neural network 140 in a forward propagation direction. Accordingly, thelearning unit 1801 obtains an output value output from the output layer 143 of the neural network 140. The dataset for learning may be stored in themain memory 92 or thestorage 93. Next, thelearning unit 1801 calculates an error between the dump body map output from the output layer 143 and the complete dump body map as the teaching data. Thelearning unit 1801 calculates an error of the weight of each connection between the neurons and an error of threshold value of each neuron by backpropagation from the calculated error of the output value. Thelearning unit 1801 updates the weight of each connection between the neurons and the threshold value of each neuron based on the calculated errors. - The
learning unit 1801 determines whether or not the output value from the complementary model M2 matches the teaching data (step S114). It may be determined that the output value matches the teaching data when an error between the output value and the teaching data is within a predetermined value. In a case where the output value from the complementary model M2 does not match the teaching data (step S114: NO), the above processing is repeated until the output value from the complementary model M2 matches the complete dump body map. As a result, the parameters of the complementary model M2 are optimized, and the complementary model M2 can be trained. - In a case where the output value from the complementary model M2 matches the teaching data (step S114: YES), the
learning unit 1801 records the complementary model M2 as a trained model including the parameters optimized by the training in the storage 93 (step S115). - As described above, according to the first embodiment, the
control device 124 specifies the three-dimensional positions of the surface of the transport object L and the bottom surface of thedump body 210 based on the captured image, and generates the dump body map indicating the distribution of the amount of the transport object L on thedump body 210 based on the three-dimensional positions. Accordingly, thecontrol device 124 can specify the distribution of the transport object L on thedump body 210. The operator can recognize the drop position of the transport object L for loading the transport object L on thedump body 210 in a well-balanced manner by recognizing the distribution of the transport object L on thedump body 210. - The
control device 124 according to the first embodiment estimates the distribution of the amount of the transport object L in the shielded part of the dump body map shielded by an obstacle. Accordingly, the operator can recognize the distribution of the amount of the transport object L even for a part of thedump body 210 that is shielded by the obstacle and cannot be captured by thestereo camera 122. - The
control device 124 according to a second embodiment specifies the distribution of the transport object L on thedump body 210 based on a type of the transport object L. -
FIG. 9 is a schematic block diagram showing a configuration of a control device according to the second embodiment. - The
control device 124 according to the second embodiment further includes atype specifying unit 1710. Thestorage 93 stores a type specifying model M3 and a plurality of complementary models M2 according to the type of the transport object L. - The
type specifying unit 1710 inputs an image of the transport object L to the type specifying model M3 to specify the type of the transport object L shown in the image. Examples of the type of transport object include clay, sand, gravel, rock, and wood. - The type specifying model M3 is realized by, for example, a trained model of deep neural network (DNN). The type specifying model M3 is a trained model which is trained, when an image showing the transport object L is input, to output the type of the transport object L. As a DNN configuration of the type specifying model M3, for example, the same or similar DNN configuration as a DNN configuration used for image recognition can be used. The type specifying model M3 is trained, for example, using a combination of an image showing the transport object L and a label representing the type of the transport object L as teaching data. The type specifying model M3 is trained using a combination of an image showing the transport object L and label data representing the type of the transport object L as the teaching data. The type specifying model M3 may be trained by transfer learning of a general trained image recognition model. The type specifying model M3 according to another embodiment may be trained by unsupervised learning or reinforcement learning.
- The
storage 93 stores the complementary model M2 for each type of the transport object L. For example, thestorage 93 stores a complementary model M2 for clay, a complementary model M2 for sand, a complementary model M2 for gravel, a complementary model M2 for rock, and a complementary model M2 for wood. Each complementary model M2 is trained, for example, using a combination of a complete dump body map generated by simulation or the like according to the type of the transport object L and an incomplete dump body map obtained by removing part of the height data from the dump body map as teaching data. -
FIG. 10 is a flowchart showing a display method of the guidance information by the control device according to the second embodiment. - First, the
data acquisition unit 1701 acquires the stereo image from the stereo camera 122 (step S21). Next, the featurepoint specifying unit 1702 inputs the right-eye image of the stereo image acquired by thedata acquisition unit 1701 to the feature point specifying model M1 stored in thestorage 93 to specify the positions of the plurality of feature points of thedump body 210 shown in the right-eye image (step S22). - The three-dimensional
data generation unit 1703 generates the point group data of the entire imaging range of thestereo camera 122 by the stereo measurement using the stereo image acquired in step S21 and the camera parameters stored in the storage 93 (step S23). - The dump
body specifying unit 1704 specifies the three-dimensional positions of the feature points based on the positions of the feature points specified in step S22 and the point group data generated in step S23 (step S24). The dumpbody specifying unit 1704 fits the dump body model VD stored in thestorage 93 to the specified positions of the feature points to specify the three-dimensional position of the bottom surface of the dump body 210 (step S25). For example, the dumpbody specifying unit 1704 disposes, based on at least three specified positions of the feature points, the dump body model VD created based on the dimensions of thedump truck 200 to be detected in a virtual space. - The
surface specifying unit 1705 extracts, from the point group data generated in step S23, a plurality of three-dimensional points in a prismatic area, which is surrounded by the front panel, the side gate, and the tailgate of thedump body 210 specified in step S25 and extends in the height direction of the front panel, to remove three-dimensional points corresponding to the background from the point group data (step S26). Thesurface specifying unit 1705 removes three-dimensional points corresponding to the position of the dump body model VD from the plurality of three-dimensional points extracted in step S6 to specify the three-dimensional position of the surface of the transport object L loaded on the dump body 210 (step S27). Thedistribution specifying unit 1706 generates the dump body map which is an elevation map with the bottom surface of thedump body 210 as a reference height, based on the plurality of three-dimensional points extracted in step S27 and the bottom surface of the dump body 210 (step S28). The dump body map may include a grid without the height data. - The
surface specifying unit 1705 specifies an area where the transport object L appears in the right-eye image based on the three-dimensional position of the surface of the transport object L specified in step S27 (step S29). For example, thesurface specifying unit 1705 specifies a plurality of pixels in the right-eye image corresponding to the plurality of three-dimensional points extracted in step S27 and determines an area composed of the plurality of specified pixels as the area where the transport object L appears. Thetype specifying unit 1710 extracts the area where the transport object L appears from the right-eye image and inputs an image related to the area to the type specifying model M3 to specify the type of the transport object L (step S30). - The
distribution estimation unit 1707 inputs the dump body map generated in step S28 to the complementary model M2 associated with the type specified in step S30 to generate the dump body map in which the height data is complemented (step S31). The guidanceinformation generation unit 1708 generates the guidance information based on the dump body map (step S32). Thedisplay control unit 1709 outputs the display signal for displaying the guidance information to the display device 125 (step S33). - As described above, according to the second embodiment, the
control device 124 estimates the distribution of the amount of the transport object L in the shielded part based on the type of the transport object L. That is, characteristics (for example, the angle of repose) of the transport object L loaded on thedump body 210 differ depending on the type of the transport object L. However, according to the third embodiment, it is possible to more accurately estimate the distribution of the transport object L in the shielded part according to the type of the transport object L. - Although embodiments have been described in detail with reference to the drawings, a specific configuration is not limited to the above, and various design changes and the like can be made.
- For example, although the
control device 124 according to the above-described embodiment is mounted on thehydraulic excavator 100, the present invention is not limited thereto. For example, thecontrol device 124 according to another embodiment may be provided in a remote server device. Thecontrol device 124 may be realized by a plurality of computers. In this case, part of the configuration of thecontrol device 124 may be provided in the remote server device. That is, thecontrol device 124 may be implemented as a transport object specifying system composed of a plurality of devices. - Although the drop target according to the above-described embodiment is the
dump body 210 of thedump truck 200, the present invention is not limited thereto. For example, the drop target according to another embodiment may be another drop target such as a hopper. - Although the captured image according to the above-described embodiment is the stereo image, the present invention is not limited thereto. For example, in another embodiment, the calculation may be performed based on one image instead of the stereo image. In this case, the
control device 124 can specify the three-dimensional position of the transport object L by using, for example, a trained model that generates depth information from the one image. - Although the
control device 124 according to the above-described embodiment complements the value of the shielded part of the dump body map by using the complementary model M2, the present invention is not limited thereto. For example, thecontrol device 124 according to another embodiment may estimate a height of the shielded part based on a rate of change or a pattern of change in the height of the transport object L near the shielded part. For example, in a case where the height of the transport object L near the shielded part becomes lower as it approaches the shielded part, thecontrol device 124 may estimate the height of the transport object L in the shielded part to a value lower than the height near the shielded part based on the rate of change in the height. - The
control device 124 according to another embodiment may estimate the height of the transport object L in the shielded part by simulation in consideration of a physical property such as the angle of repose of the transport object L. Thecontrol device 124 according to another embodiment may deterministically estimate the height of the transport object L in the shielded part based on cellular automaton in which each grid of the dump body map is regarded as a cell. - The
control device 124 according to another embodiment may not complement the dump body map and may display information related to the dump body map including a part where the height data is missing. -
FIG. 11A is a diagram showing a first example of a method for calculating an amount of a transport object in a dump body.FIG. 11B is a diagram showing a second example of the method for calculating the amount of the transport object in the dump body. - As shown in
FIG. 11A , the dump body map according to the above-described embodiment is represented by a height from a bottom surface L1 of thedump body 210 to an upper limit of the loading on thedump body 210, but the present invention is not limited thereto. - For example, as shown in
FIG. 11B , the dump body map according to another embodiment may represent a height from another reference plane L3 with respect to the bottom surface to a surface L2 of the transport object L. In the example shown inFIG. 11B , the reference plane L3 is a plane parallel to the ground surface and passing through a point of the bottom surface closest to the ground surface. In this case, the operator can easily recognize the amount of the transport object L until thedump body 210 is full, regardless of an inclination of thedump body 210. - Although the
control device 124 according to the above-described embodiment generates the dump body map based on the bottom surface of thedump body 210 and the surface of the transport object L, the present invention is not limited thereto. For example, thecontrol device 124 according to another embodiment may calculate the dump body map based on an opening surface of thedump body 210, the surface of the transport object, and a height from the bottom surface to the opening surface of thedump body 210. That is, thecontrol device 124 may calculate the dump body map by subtracting, from the height from the bottom surface to the opening surface of thedump body 210, a distance from an upper end surface of the dump body to the surface of the transport object L. The dump body map according to another embodiment may be based on the opening surface of thedump body 210. - Although the guidance
information generation unit 1708 according to the above-described embodiment extracts the feature points from the right-eye image using the feature point specifying model M1, the present invention is not limited thereto. For example, in another embodiment, the guidanceinformation generation unit 1708 may extract the feature points from the left-eye image using the feature point specifying model M1. - The transport object specifying device according to the present invention can specify the distribution of the transport object in the drop target.
Claims (14)
1. A transport object specifying device of a work machine, the transport object specifying device comprising:
an image acquisition unit that acquires a captured image showing a drop target of the work machine in which a transport object is dropped;
a drop target specifying unit that specifies a three-dimensional position of at least part of the drop target based on the captured image;
a three-dimensional data generation unit that generates depth data, which is three-dimensional data representing a depth of the captured image, based on the captured image; and
a surface specifying unit that specifies a three-dimensional position of a surface of the transport object in the drop target by removing, from the depth data, a part corresponding to the drop target based on the three-dimensional position of the at least part of the drop target.
2. The transport object specifying device according to claim 1 , further comprising:
a feature point specifying unit that specifies a position of a feature point of the drop target based on the captured image,
the drop target specifying unit specifying the three-dimensional position of the at least part of the drop target based on the position of the feature point.
3. The transport object specifying device according to claim 1 , wherein
the drop target specifying unit specifies the three-dimensional position of the at least part of the drop target based on the captured image and a target model, which is a three-dimensional model indicating a shape of the drop target.
4. The transport object specifying device according to claim 1 , wherein
the surface specifying unit
extracts, from the depth data, a three-dimensional position in a prismatic area which, is surrounded by a wall portion of the drop target and extends in a height direction of the wall portion, and
removes the part corresponding to the drop target in the extracted three-dimensional position to specify the three-dimensional position of the surface of the transport object.
5. The transport object specifying device according to claim 1 , further comprising:
a distribution specifying unit that generates distribution information indicating a distribution of an amount of the transport object in the drop target based on
the three-dimensional position of the surface of the transport object in the drop target and
the three-dimensional position of the at least part of the drop target.
6. The transport object specifying device according to claim 5 , further comprising:
a distribution estimation unit that estimates a distribution of an amount of the transport object in a shielded part of the distribution information shielded by an obstacle.
7. The transport object specifying device according to claim 6 , wherein
the distribution estimation unit inputs the distribution information generated by the distribution specifying unit to a complementary model to generate distribution information that complements a value of the shielded part, and
the complementary model is a trained model, which, when distribution information with some missing values is input, outputs distribution information that complements the missing values.
8. The transport object specifying device according to claim 6 , wherein
the distribution estimation unit generates distribution information that complements a value of the shielded part based on a rate of change or a pattern of change in a three-dimensional position of the transport object near the shielded part.
9. The transport object specifying device according to claim 6 , wherein
the distribution estimation unit estimates a distribution of an amount of the transport object in the shielded part based on a type of the transport object.
10. The transport object specifying device according to claim 1 , wherein
the captured image is a stereo image including at least a first image and a second image captured by a stereo camera.
11. A work machine including the transport object specifying device according to claim 1 , the work machine further comprising:
work equipment usable to transport the transport object;
an imaging device; and
a display device that displays information about the transport object in the drop target specified by the transport object specifying device.
12. A transport object specifying method of a work machine, comprising:
acquiring a captured image showing a drop target of the work machine in which a transport object is dropped;
specifying a three-dimensional position of at least part of the drop target based on the captured image;
generating depth data, which is three-dimensional data representing a depth of the captured image, based on the captured image; and
removing, from the depth data, a part corresponding to the drop target based on the three-dimensional position of the at least part of the drop target to specify a three-dimensional position of a surface of the transport object in the drop target.
13. A method for producing a complementary model, which, when distribution information with some missing values is input, outputs distribution information that complements the missing values, the method comprising:
acquiring distribution information indicating a distribution of an amount of a transport object in a drop target of a work machine and incomplete distribution information in which some values of the distribution information are missing, as a dataset for learning; and
training the complementary model using the dataset for learning such that when the incomplete distribution information is used as an input value, the distribution information becomes an output value.
14. (canceled)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2018-163671 | 2018-08-31 | ||
JP2018163671A JP7311250B2 (en) | 2018-08-31 | 2018-08-31 | Device for identifying goods carried by working machine, working machine, method for identifying goods carried by working machine, method for producing complementary model, and data set for learning |
PCT/JP2019/028454 WO2020044848A1 (en) | 2018-08-31 | 2019-07-19 | Device to specify cargo carried by construction machinery, construction machinery, method to specify cargo carried by construction machinery, method for producing interpolation model, and dataset for learning |
Publications (1)
Publication Number | Publication Date |
---|---|
US20210272315A1 true US20210272315A1 (en) | 2021-09-02 |
Family
ID=69645231
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/260,069 Pending US20210272315A1 (en) | 2018-08-31 | 2019-07-19 | Transport object specifying device of work machine, work machine, transport object specifying method of work machine, method for producing complementary model, and dataset for learning |
Country Status (5)
Country | Link |
---|---|
US (1) | US20210272315A1 (en) |
JP (1) | JP7311250B2 (en) |
CN (1) | CN112513563B (en) |
DE (1) | DE112019003049T5 (en) |
WO (1) | WO2020044848A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220195690A1 (en) * | 2019-09-18 | 2022-06-23 | Sumitomo Heavy Industries, Ltd. | Shovel |
US11414837B2 (en) * | 2018-08-31 | 2022-08-16 | Komatsu Ltd. | Image processing system, display device, image processing method, method for generating trained model, and dataset for learning |
US20220364873A1 (en) * | 2021-05-12 | 2022-11-17 | Deere & Company | System and method for assisted positioning of transport vehicles for material discharge in a worksite |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7374142B2 (en) * | 2021-03-01 | 2023-11-06 | 日立建機株式会社 | wheel loader |
US11965308B2 (en) | 2021-05-12 | 2024-04-23 | Deere & Company | System and method of truck loading assistance for work machines |
JP2023088646A (en) * | 2021-12-15 | 2023-06-27 | 株式会社小松製作所 | Method for calculating repose angle of excavated matter held in bucket, system for calculating repose angle of excavated matter held in bucket, and loading machine |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120114181A1 (en) * | 2010-11-01 | 2012-05-10 | Borthwick James R | Vehicle pose estimation and load profiling |
WO2018043301A1 (en) * | 2016-09-02 | 2018-03-08 | 株式会社小松製作所 | Work machine graphics display system |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH01156606A (en) * | 1987-12-15 | 1989-06-20 | Matsushita Electric Works Ltd | Optical interference type shape measuring instrument |
JP3091648B2 (en) * | 1994-09-26 | 2000-09-25 | 沖電気工業株式会社 | Learning Hidden Markov Model |
JPH11211438A (en) * | 1998-01-22 | 1999-08-06 | Komatsu Ltd | Load carrying platform load volume measuring device |
JP3038474B2 (en) * | 1998-09-11 | 2000-05-08 | 五洋建設株式会社 | Method and apparatus for measuring the amount of soil loaded on an earth moving ship |
JP2004061300A (en) * | 2002-07-29 | 2004-02-26 | Asahi Shoji Ltd | Laser type angle detection device, deflection measuring device of crank shaft, deflection measuring method of crank shaft, and crank shaft |
JP2005220633A (en) | 2004-02-06 | 2005-08-18 | Ohbayashi Corp | Device and method for detecting conveyance soil and sand amount of belt conveyor |
JP5835881B2 (en) * | 2010-10-06 | 2015-12-24 | 株式会社東芝 | Medical image processing apparatus and medical image processing program |
CN104200657B (en) * | 2014-07-22 | 2018-04-10 | 杭州智诚惠通科技有限公司 | A kind of traffic flow parameter acquisition method based on video and sensor |
US10208459B2 (en) * | 2014-12-12 | 2019-02-19 | Hitachi, Ltd. | Volume estimation device and work machine using same |
JP6567940B2 (en) | 2015-10-05 | 2019-08-28 | 株式会社小松製作所 | Construction management system |
JP6674846B2 (en) | 2016-05-31 | 2020-04-01 | 株式会社小松製作所 | Shape measuring system, work machine and shape measuring method |
CN106839977B (en) * | 2016-12-23 | 2019-05-07 | 西安科技大学 | Shield dregs volume method for real-time measurement based on optical grating projection binocular imaging technology |
CN106885531B (en) * | 2017-04-20 | 2018-12-18 | 河北科技大学 | Wagon box based on two-dimensional laser radar describes device 3 D scanning system scaling method |
CN107168324B (en) * | 2017-06-08 | 2020-06-05 | 中国矿业大学 | Robot path planning method based on ANFIS fuzzy neural network |
CN108332682A (en) * | 2018-02-06 | 2018-07-27 | 黑龙江强粮安装饰工程有限公司 | Novel granary dynamic storage unit weight monitoring system and monitoring method |
-
2018
- 2018-08-31 JP JP2018163671A patent/JP7311250B2/en active Active
-
2019
- 2019-07-19 CN CN201980050449.XA patent/CN112513563B/en active Active
- 2019-07-19 DE DE112019003049.3T patent/DE112019003049T5/en active Pending
- 2019-07-19 US US17/260,069 patent/US20210272315A1/en active Pending
- 2019-07-19 WO PCT/JP2019/028454 patent/WO2020044848A1/en active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120114181A1 (en) * | 2010-11-01 | 2012-05-10 | Borthwick James R | Vehicle pose estimation and load profiling |
WO2018043301A1 (en) * | 2016-09-02 | 2018-03-08 | 株式会社小松製作所 | Work machine graphics display system |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11414837B2 (en) * | 2018-08-31 | 2022-08-16 | Komatsu Ltd. | Image processing system, display device, image processing method, method for generating trained model, and dataset for learning |
US20220195690A1 (en) * | 2019-09-18 | 2022-06-23 | Sumitomo Heavy Industries, Ltd. | Shovel |
US20220364873A1 (en) * | 2021-05-12 | 2022-11-17 | Deere & Company | System and method for assisted positioning of transport vehicles for material discharge in a worksite |
US11953337B2 (en) * | 2021-05-12 | 2024-04-09 | Deere & Company | System and method for assisted positioning of transport vehicles for material discharge in a worksite |
Also Published As
Publication number | Publication date |
---|---|
CN112513563B (en) | 2023-01-13 |
WO2020044848A1 (en) | 2020-03-05 |
CN112513563A (en) | 2021-03-16 |
JP2020034527A (en) | 2020-03-05 |
DE112019003049T5 (en) | 2021-03-11 |
JP7311250B2 (en) | 2023-07-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11414837B2 (en) | Image processing system, display device, image processing method, method for generating trained model, and dataset for learning | |
US20210272315A1 (en) | Transport object specifying device of work machine, work machine, transport object specifying method of work machine, method for producing complementary model, and dataset for learning | |
US12094151B2 (en) | Image processing system, image processing method, learned model generation method, and data set for learning | |
US11417008B2 (en) | Estimating a volume of contents in a container of a work vehicle | |
US11494930B2 (en) | Techniques for volumetric estimation | |
CN110805093B (en) | Container angle sensing with feedback loop control using vision sensors | |
Se et al. | Vision based modeling and localization for planetary exploration rovers | |
US11591776B2 (en) | Earth-moving machine sensing and control system | |
US8903689B2 (en) | Autonomous loading | |
US9990543B2 (en) | Vehicle exterior moving object detection system | |
KR20190120322A (en) | Method, system, method for manufacturing trained classification model, training data, and method for manufacturing training data | |
Guevara et al. | Point cloud-based estimation of effective payload volume for earthmoving loaders | |
JP2014228941A (en) | Measurement device for three-dimensional surface shape of ground surface, runnable region detection device and construction machine mounted with the same, and runnable region detection method | |
CN114170596A (en) | Posture recognition method and device, electronic equipment, engineering machinery and storage medium | |
Mandici et al. | Aggregate road surface based environment representation using digital elevation maps | |
JP7175245B2 (en) | working machine | |
JP7501481B2 (en) | Distance estimation device, distance estimation method, and computer program for distance estimation | |
CN118269993A (en) | Vehicle control method, device, equipment, storage medium, product and system | |
CN116189141A (en) | Determination method of target information and computer equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |