US20200184667A1 - System for Volume Dimensioning Via Holographic Sensor Fusion - Google Patents
System for Volume Dimensioning Via Holographic Sensor Fusion Download PDFInfo
- Publication number
- US20200184667A1 US20200184667A1 US16/786,268 US202016786268A US2020184667A1 US 20200184667 A1 US20200184667 A1 US 20200184667A1 US 202016786268 A US202016786268 A US 202016786268A US 2020184667 A1 US2020184667 A1 US 2020184667A1
- Authority
- US
- United States
- Prior art keywords
- target object
- display
- processor
- volume dimensioning
- mobile device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1408—Methods for optical code recognition the method being specifically adapted for the type of code
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/45—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
-
- H04N5/2258—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/445—Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
- H04N5/44504—Circuit details of the additional information generator, e.g. details of the character or graphics signal generator, overlay mixing circuits
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/04—Indexing scheme for image data processing or generation, in general involving 3D image data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/08—Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
Definitions
- the present application is related to and claims the benefit of the earliest available effective filing dates from the following listed applications (the “Related Applications”) (e.g., claims earliest available priority dates for other than provisional patent applications (e.g., under 35 USC ⁇ 120 as a continuation in part) or claims benefits under 35 USC ⁇ 119(e) for provisional patent applications, for any and all parent, grandparent, great-grandparent, etc. applications of the Related Applications).
- the “Related Applications” e.g., claims earliest available priority dates for other than provisional patent applications (e.g., under 35 USC ⁇ 120 as a continuation in part) or claims benefits under 35 USC ⁇ 119(e) for provisional patent applications, for any and all parent, grandparent, great-grandparent, etc. applications of the Related Applications).
- LIDAR light detection and ranging
- a wearable augmented reality (AR) device e.g., virtual reality (VR), mixed reality (MR) device
- the wearable device communicates with a system for volume dimensioning via two-dimensional (2D)/three-dimensional (3D) sensor fusion.
- the AR device display to the user a two-dimensional (2D) image stream of a field of view (FOV) including target objects and 3D imaging data of the FOV including point clouds of each target object within the FOV.
- Target objects are identified by correlating the 2D and 3D image streams and generating a holographic model of the target object overlaid on the video stream, with adjustable surface, edge, and vertex guides.
- Precise dimensions and unique object identifiers of the target object are determined and decoded and displayed to the user via the wearable device; the user may manipulate or adjust the display via gestures or commands detectable by the wearable device.
- FIG. 1 is a block diagram illustrating a system for volume dimensioning via sensor fusion, in accordance with example embodiments of this disclosure
- FIGS. 2A and 2B are diagrammatic illustrations of a mobile device of the volume dimensioning system of FIG. 1 ;
- FIGS. 3A and 3B are diagrammatic illustrations of a mobile device of the volume dimensioning system of FIG. 1 ;
- FIGS. 4A through 4C are diagrammatic illustrations of a mobile device of the volume dimensioning system of FIG. 1 ;
- FIGS. 5A and 5B are diagrammatic illustrations of a mobile device of the volume dimensioning system of FIG. 1 ;
- FIGS. 6A and 6B illustrate operations of, and a mobile device of, the volume dimensioning system of FIG. 1 ;
- FIG. 7 is a diagrammatic illustration of a mobile device of the volume dimensioning system of FIG. 1 ;
- FIGS. 8A and 8B are diagrammatic illustrations of an augmented reality (AR) device of the volume dimensioning system of FIG. 1 ;
- AR augmented reality
- FIGS. 9A and 9B illustrate operations of, and the mobile device of, the volume dimensioning system of FIG. 1 ;
- FIGS. 10A and 10B illustrate operations of, and the mobile device of, the volume dimensioning system of FIG. 1 ;
- FIG. 11 is a flow diagram illustrating a method for volume dimensioning via sensor fusion, in accordance with example embodiments of this disclosure.
- a letter following a reference numeral is intended to reference an embodiment of the feature or element that may be similar, but not necessarily identical, to a previously described element or feature bearing the same reference numeral (e.g., 1 , 1 a, 1 b ).
- Such shorthand notations are used for purposes of convenience only and should not be construed to limit the disclosure in any way unless expressly stated to the contrary.
- any reference to “one embodiment” or “some embodiments” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment disclosed herein.
- the appearances of the phrase “in some embodiments” in various places in the specification are not necessarily all referring to the same embodiment, and embodiments may include one or more of the features expressly described or inherently present herein, or any combination of sub-combination of two or more such features, along with any other features which may not necessarily be expressly described or inherently present in the instant disclosure.
- the volume dimensioning system includes two-dimensional (2D) and three-dimensional (3D) image sensors incorporated into or attached to a mobile device, e.g., a smartphone, tablet, phablet, or like portable processor-enabled device.
- the 2D imager captures an image stream or sequence (e.g., streaming video at 30 or 60 fps) corresponding to a field of view of the device, while the 3D imager creates a point cloud or similar point set (e.g., a 3D mesh representation) of the field of view.
- the image sources can be fused to detect a target object within the field of view (e.g., a shipping container or like cuboid object traveling through a supply chain). By analyzing the fused imagery, the target object can be accurately measured and any encoded information relevant to the object decoded and updated
- the volume dimensioning system 100 may include a mobile device 102 capable of being carried by an operator or user, the mobile device 102 including 2D image sensors 104 (e.g., a visible-light camera), 3D image sensors 106 , image and control processors 108 , a touch-sensitive display surface 110 , and a wireless transceiver 112 .
- 2D image sensors 104 e.g., a visible-light camera
- 3D image sensors 106 e.g., a visible-light camera
- image and control processors 108 e.g., a touch-sensitive display surface 110
- a wireless transceiver 112 e.g., a wireless transceiver
- the mobile device 102 may additionally include a clock 114 or time sensor, a Global Positioning System (GPS) receiver 116 or similar position sensor for determining a current position of the mobile device, and an inertial measurement unit 118 (IMU) or similar inertial sensor (e.g., accelerometer, magnetometer, gyrometer, compass) for determining a current orientation of the mobile device (or for tracking the orientation, and the rate of change thereof, over time).
- GPS Global Positioning System
- IMU inertial measurement unit 118
- the volume dimensioning system 100 may incorporate IMUs integrated into the 2D image sensors 104 or into the 3D image sensors 106 .
- the 3D image sensors 106 may include imaging systems including infrared illuminators combined with multiple embedded cameras (e.g., Intel RealSense or other like triangulating systems), laser-based light detection and ranging (LIDAR) systems incorporating onboard photodetectors to track reflected beams, time of flight (ToF) camera systems, or any other like sensor system capable of producing 3D spatial information of proximate objects.
- imaging systems including infrared illuminators combined with multiple embedded cameras (e.g., Intel RealSense or other like triangulating systems), laser-based light detection and ranging (LIDAR) systems incorporating onboard photodetectors to track reflected beams, time of flight (ToF) camera systems, or any other like sensor system capable of producing 3D spatial information of proximate objects.
- LIDAR laser-based light detection and ranging
- ToF time of flight
- the 3D image sensors 106 may incorporate inertial or orientation sensors or combinations thereof, e.g., accelerometers, gyroscope
- the mobile device 102 may be oriented toward a target object 120 in such a way that the 2D image sensors 104 and 3D image sensors 106 simultaneously capture image data from a common field of view in which the target object 120 is situated.
- the target object 120 may include a shipping box or container currently traveling through a supply chain, e.g., from a known origin to a known destination.
- the target object 120 may be freestanding on a floor, table, or other flat surface; in some embodiments the target object may be secured to a pallet 122 or similar structural foundation, either individually or in a group of such objects, for storage or transport.
- the target object 120 is preferably substantially cuboid in shape, e.g., having six rectangular surfaces ( 120 a - c ) intersecting at right angles.
- the target object 120 may not itself be perfectly cuboid but may fit perfectly within a minimum cuboid volume 124 of determinable dimensions (e.g., the minimum cuboid volume necessary to fully surround or encompass the target object).
- the precise dimensions of the target object 120 may be unknown to the operator of the mobile device 102 but determinable via sensor fusion of the respective sensor inputs 126 , 128 of the 2D imager 104 and 3D imager 106 .
- the volume dimensioning system 100 may detect the target object 120 via a single sensor array, e.g., either the 2D image sensors 104 or the 3D image sensors 106 , as opposed to a combination of 2D and 3D image sensors fusing their respective data outputs to aid detection. Similarly, the volume dimensioning system 100 may simultaneously detect and analyze multiple target objects 120 within a particular field of view.
- the wireless transceiver 112 may enable the establishment of wireless links to remote sources, e.g., physical servers 130 and cloud-based storage 132 .
- the wireless transceiver 112 may establish a wireless link 112 a to a remote operator 134 situated at a physical distance from the mobile device 102 and the target object 120 , such that the remote operator may visually interact with the target object 120 and submit control input to the mobile device 102 .
- the wireless transceiver 112 may establish a wireless link 112 a to an augmented reality (AR) viewing device 136 (e.g., a virtual reality (VR) or mixed reality (MR) device worn on the head of a viewer, or proximate to the viewer's eyes, and capable of displaying to the viewer real-world objects and environments, synthetic objects and environments, or combinations thereof).
- AR augmented reality
- the AR viewing device 136 may allow the user to interact with the target object 120 and/or the mobile device 102 (e.g., submitting control input to manipulate the field of view, or a representation of the target object situated therein) via physical, ocular, or aural control input detected by the AR viewing device.
- the mobile device 102 may include a memory 138 or other like means of data storage accessible to the image and control processors 108 , the memory capable of storing reference data accessible to the volume dimensioning system 100 to make additional determinations with respect to the target object 120 .
- the mobile device 102 may include a microphone 140 for receiving aural control input from the user/operator, e.g., verbal commands to the volume dimensioning system 100 .
- the mobile device 102 a may be implemented and may function similarly to the mobile device 102 of FIG. 1 , except that the mobile device 102 may be oriented toward the target object 120 in such a way that the 2D imager ( 104 , FIG. 1 ) captures an image stream ( 126 , FIG. 1 ) of a field of view including the target object (or multiple objects).
- the display surface 110 may display the image stream 126 to the operator of the mobile device 102 in real time or near real time.
- the operator of the mobile device 102 may position or orient the target object 120 (or the mobile device relative to the target object) to optimally position the target object within the field of view for image analysis.
- the target object 120 may be positioned proximate to a floor surface ( 202 ) and/or wall ( 204 ) of optimal contrast to the target object (e.g., in color, in texture) or lit so as to optimize the contrast between the target object and the floor/wall.
- a floor surface ( 202 ) and/or wall ( 204 ) of optimal contrast to the target object e.g., in color, in texture
- lit e.g., in color, in texture
- the operator may position and orient the mobile device 102 relative to the target object 120 so that the entire upper surface ( 206 ) of the target object is clearly visible to the 2D imager 104 , and that a forward vertical edge ( 208 ) and forward top corner or vertex ( 210 ) faces the 2D imager.
- the volume dimensioning system ( 100 , FIG. 1 ) may analyze the target object 120 based on cumulative views captured from multiple orientations around the target object 120 .
- the image and control processors may analyze the image stream 126 or component frames thereof to distinguish the target object 120 from its surroundings. For example, the image and control processors 108 may attempt to identify patterns and shifts in color and texture that may be indicative of the target object 120 or its parameters (e.g., surfaces, edges, vertices).
- the mobile device 102 b may be implemented and may function similarly to the mobile device 102 a of FIG. 2A , except that the mobile device 102 b may display via the display surface 110 , in addition to the image stream ( 126 , FIG. 1 ) captured by the 2D imager ( 104 , FIG. 1 ), 3D image data ( 128 , FIG. 1 ) captured by the 3D imager ( 106 , FIG. 1 ).
- the field of view of the 3D imager 106 may substantially correspond to the field of view of the image stream 126 captured by the 2D imager 104 .
- the 3D image data 128 may include a stream of pixel sets, each pixel set substantially corresponding to a frame of the 2D image stream 126 .
- the pixel set may include a point cloud 212 (e.g., point map) substantially corresponding to the target object 120 .
- Each point 214 of the point cloud 212 may include a coordinate set (e.g., XY) locating the point relative to the field of view (e.g., to the frame, to the pixel set) as well as plane angle and depth data of the point, e.g., the distance of the point from the mobile device 102
- a coordinate set e.g., XY
- plane angle and depth data of the point e.g., the distance of the point from the mobile device 102
- the 3D imager 106 may analyze depth information about the target object 120 and its environment as shown within its field of view. For example, the 3D imager 106 may identify the floor 202 as a plane of gradually increasing depth that meets an intersecting plane (e.g., a rear wall 204 or a candidate plane surface 216 corresponding to a surface ( 120 a - c , FIG. 1 ) of the target object 120 ). Based on the intersections of the candidate plane surfaces 216 of the target object 120 (e.g., with each other or with the floor 202 ), the 3D imager 106 may identify candidate edges 218 . Similarly, the intersection of three plane surfaces, or the intersection of two candidate edges 218 , may indicate a candidate vertex 220 .
- an intersecting plane e.g., a rear wall 204 or a candidate plane surface 216 corresponding to a surface ( 120 a - c , FIG. 1 ) of the target object 120 .
- the 3D imager 106 may identify candidate edges 2
- the volume dimensioning system 100 may account for imperfect data sets, e.g., gaps or holes in the point cloud, via plane identification.
- the volume dimensioning system may analyze 3D spatial information 128 (independently or fused with the 2D image stream 126 ) to infer the planes of the target object 120 , e.g., on the basis of a sufficient number of identified points aligned in a plane or nearly enough aligned (e.g., within a predetermined range) to derive the existence of a plane.
- the volume dimensioning system 100 may identify the target object 120 and its component planes quickly enough, or to a sufficient level of confidence, that fusion of 2D image data 126 may not be necessary for optimal performance. In some embodiments, the volume dimensioning system may similarly infer the edges or vertices of the target object 120 based on the placement or alignment of individual points.
- the mobile device 102 c may be implemented and may function similarly to the mobile device 102 b of FIG. 2B , except that with respect to the mobile device 102 c, the volume dimensioning system ( 100 , FIG. 1 ) may correlate the 2D image stream ( 126 , FIG. 1 ) and 3D point cloud ( 212 , FIG. 2B ) to identify the target object 120 and generate a holographic model 302 corresponding to the target object.
- the volume dimensioning system 100 may identify candidate parameters of the target object 120 and digitally represent these candidate parameters in the holographic model 302 .
- the volume dimensioning system 100 may tentatively identify candidate surfaces 304 , candidate edges 306 , and candidate vertices 308 of the holographic model 302 , which candidate surfaces, edges, and vertices correspond to the real-world parameters (e.g., surfaces 120 a - c , FIG. 1 ) of the target object 120 .
- the holographic model 302 may be displayed to the operator of the mobile device 102 c (e.g., via the display surface 110 ) as a three-dimensional object.
- the manner or style in which the holographic model 302 is displayed may vary based on operator preference or control input, e.g., wireframe, semitransparent solid, monochrome, multicolor.
- the volume dimensioning system 100 will generate the holographic model 302 by overlaying the the target object 120 with edge lines 310 (e.g., edge guides), vertex points 312 (e.g., vertex guides), and/or surface guides 314 (e.g., solid, shaded, textured planes, or planes of varying opaqueness) as the volume dimensioning system locks onto the target object 120 , indicating to the user/operator that the target object has been positively identified.
- edge lines 310 e.g., edge guides
- vertex points 312 e.g., vertex guides
- surface guides 314 e.g., solid, shaded, textured planes, or planes of varying opaqueness
- the 2D image of the target object 120 may be transposed onto or otherwise incorporated into the holographic model 302 , such that the 3D holographic model 302 appears substantially similar to the user/operator as would the real target object 120 (e.g., when viewed directly or via the 2D imager 104 ).
- the volume dimensioning system 100 may be trained via machine learning to recognize and lock onto a target object 120 , positively identifying the target object and distinguishing the target object from its surrounding environment (e.g., the field of view of the 2D imager 104 and 3D imager 106 including the target object as well as other candidate objects, which may additionally be locked onto as target objects and dimensioned).
- the volume dimensioning system 100 may include a recognition engine trained on positive and negative images of a particular object specific to a desired use case.
- the recognition engine may be trained to specific latitudes, longitudes, and locations, such that the performance of the recognition engine may be driven in part by the current location of the mobile device 102 a, the current time of day, the current time of year, or some combination thereof.
- the mobile device 102 d may be implemented and may function similarly to the mobile device 102 c of FIG. 3A , except that with respect to the mobile device 102 d, once the holographic model 302 is generated by the volume dimensioning system ( 100 , FIG. 1 ), the operator may manipulate the holographic model 302 as displayed by the display surface 110 . For example, by sliding his/her finger ( 316 ) across the touch-sensitive display surface 110 , the operator may move ( 318 ) the holographic model 302 relative to the display surface (e.g., and relative to the image stream 126 and target object (or multiple objects simultaneously in the field of view) 102 portrayed thereby) or rotate ( 320 ) the holographic model.
- the display surface e.g., and relative to the image stream 126 and target object (or multiple objects simultaneously in the field of view) 102 portrayed thereby
- the candidate parameters of the holographic model 302 may be shifted, resized, or corrected as shown below.
- the holographic model 302 may be manipulated based on aural control input 322 submitted by the operator (e.g., via a microphone ( 140 , FIG. 1 ) disposed within the mobile device 102 d ).
- the volume dimensioning system 100 may respond to verbal commands from the operator, e.g., to shift or rotate the holographic model 302 , or to select and shift an edge guide 310 , vertex guide 312 , or surface guide 314 .
- the mobile device 102 e may be implemented and may function similarly to the mobile device 102 d of FIG. 3B , except that with respect to the mobile device 102 e, when the target object ( 120 , FIG. 3B ) has been identified (e.g., the 3D point cloud ( 212 , FIG. 2B ) has been fused and correlated with the 2D image stream ( 126 , FIG. 1 ) of the target object ( 120 ) and the candidate edges ( 306 , FIG. 3B ) and other candidate parameters of the holographic model 302 have been identified, the volume dimensioning system ( 100 , FIG. 1 ) may measure the holographic model 302 to determine the dimensions of the target object 120 to which the holographic model 302 corresponds.
- the target object 120 , FIG. 3B
- the 3D point cloud 212 , FIG. 2B
- the candidate edges 306 , FIG. 3B
- the volume dimensioning system 100 , FIG. 1
- the volume dimensioning system may measure the holographic model
- the holographic model 302 of the cuboid target object 120 may itself correspond to a cuboid digital representation.
- a top surface 304 a, a forward edge 306 a, and a top forward corner or vertex 308 may all be displayed via the display surface 110 .
- nine of the twelve edges of the holographic model 302 may be displayed, or three edges for each linear dimension of the holographic model.
- the volume dimensioning system 100 may determine the precise dimensions ( 402 ) of the target object 120 (e.g., x-axis width, y-axis breadth, z-axis height, volume) by measuring edges of the holographic model 302 corresponding thereto. For example, with respect to the z-axis height ( 404 ) of the target object 120 , the volume dimensioning system 100 may measure the vertical edges 306 a - c of the holographic model 302 .
- the precise dimensions ( 402 ) of the target object 120 e.g., x-axis width, y-axis breadth, z-axis height, volume
- the volume dimensioning system 100 may measure the vertical edges 306 a - c of the holographic model 302 .
- the volume dimensioning system 100 may account for general model or technology variations, errors, or holes (e.g., incompletions, gaps) in the 3D point cloud 212 which may skew individual edge measurements (particularly if the hole coincides with a vertex 308 (e.g., an endpoint of the edge) of the holographic model 302 .
- the volume dimensioning system 100 would apply the same methodology in measuring the other parallel edges ( 306 d - f , 306 g - i ; FIGS. 4B and 4C ) to produce a line length of all three dimensions 406 , 408 , 410 of the target object.
- the volume dimensioning system 100 may adjust the measuring process (e.g., based on control input from the operator) for increased accuracy or speed.
- the measurement of a given dimension may be based on multiple readings or pollings of the holographic model 302 (e.g., by generating multiple holographic models per second on a frame-by-frame basis and selecting “good” measurements to generate a result set (e.g., 10 measurement sets) for averaging).
- the three measurements of the vertical edges 306 a - c may be averaged to determine a given dimension.
- the measurement may be counted as a “good” reading for purposes of inclusion within a result set.
- the confirmation tolerance may be increased by requiring all three edges 306 a - c to be within the threshold variance for inclusion in the result set.
- the x-axis width 406 , y-axis breadth 408 , and volume 410 of the holographic model 302 may similarly be determined by analysis and measurement of the x-axis edges 306 d - f and y-axis edges 306 g - i of the holographic model 302 .
- the volume dimensioning system 100 may proceed at a reduced confidence level if measurements cannot be established at full confidence.
- the exterior surface of the target object 120 may be matte-finished, light-absorbing, or otherwise treated in such a way that the volume dimensioning system (particularly with respect to the 2D imager 104 ) may have difficulty accurately determining or measuring surfaces, edges, and vertices of the holographic model 302 .
- the volume dimensioning system 100 may, for example, reduce the number of minimum confirmations required for an acceptable measure (e.g., from 3 to 2) or analyze additional frames per second (e.g., sacrificing operational speed for enhanced accuracy).
- the confidence condition level ( 412 ) may be displayed to the user and stored in the dataset corresponding to the target object 120 .
- the mobile device 102 f may be implemented and may function similarly to the mobile device 102 e of FIGS. 4A through 4C , except that with respect to the mobile device 102 f, the volume dimensioning system ( 100 , FIG. 1 ) may account for any pallets 122 or other foundations to which the target object 120 is attached, determining the minimum possible dimensions 124 of the target object (e.g., based on the minimum possible amount of shelf space the target object attached to the pallet would occupy in a vehicle, in a warehouse, or elsewhere in the supply chain.
- the holographic model 302 a may be implemented and may function similarly to the holographic model 302 of FIGS. 4A-B , except that the holographic model 302 a may account for the pallet 122 in determining the minimum possible edges ( 502 ), minimum possible surfaces ( 504 ), and minimum possible volume of the target object 120 attached to the pallet 122 .
- the volume dimensioning system 100 a may be implemented and may function similarly to the volume dimensioning system 100 of FIG. 1 , except that the volume dimensioning system 100 a may monitor the onboard IMU ( 118 , FIG. 1 ) of the mobile device 102 (e.g., or inertial/orientation sensors integrated into the 2D imager ( 104 , FIG. 1 ) or 3D imager ( 106 , FIG. 1 ) to detect difficulty in the identification of candidate surfaces ( 304 , FIGS. 3A-B ), candidate edges ( 306 , FIGS. 3A-B ), and candidate vertices ( 306 , FIGS. 3A-B ) from the holographic model 302 ( FIGS. 3A-B ) corresponding to the target object 120 .
- the onboard IMU 118 , FIG. 1
- the mobile device 102 e.g., or inertial/orientation sensors integrated into the 2D imager ( 104 , FIG. 1 ) or 3D imager ( 106
- the IMU 118 may detect excessive shifts in the orientation of the mobile device 102 as the user ( 602 ) moves the mobile device around ( 604 ) and the volume dimensioning system 100 a attempts to lock into the parameters of the target object via the 2D image stream 126 and the 3D image data 128 . Similarly, the IMU 118 may notice rotational movement by the user 604 around the target object 120 and take this movement into account in the generation of the 3D holographic model 302 .
- the mobile device 102 g may be implemented and may function similarly to the mobile device 102 of FIG. 6A , except that with respect to the mobile device 102 g, the volume dimensioning system 100 a may respond to the detection of certain types of motion by the IMU 118 by displaying surface guides 314 (e.g., planar guides), edge guides 310 , and vertex guides 312 to assist the user 602 in manually adjusting the candidate planes 310 , edges 306 and candidate vertices 308 of the holographic model 302 .
- surface guides 314 e.g., planar guides
- edge guides 310 e.g., edge guides 310
- vertex guides 312 e.g., vertex guides
- the user 602 may manipulate the edge guides 310 , surface guides 314 , and/or vertex guides 312 (e.g., via the touch-sensitive display surface 110 ) to precisely set the candidate edges 306 and candidate vertices 308 so that the volume dimensioning system 100 a may accurately measure the holographic model 302 .
- the volume dimensioning system 100 a may sense a user/operator manipulating a surface guide 314 , edge guide 310 , or vertex guide 312 and understand intuitively where the user is intending to place the vertex guide ( 312 a ), in which case the vertex guide 312 would automatically snap into the nearest logical place.
- the mobile device 102 h may be implemented and may function similarly to the mobile device 102 g of FIG. 6B and the mobile devices 102 , 102 a - f of FIGS. 1 through 5B , except that with respect to the mobile device 102 h, the volume dimensioning system ( 100 , FIG. 1 ; 100 a , FIG. 6A ) may analyze the 2D image stream ( 126 , FIG. 1 ) to detect and decode any encoded information on an exterior surface of the target object ( 120 , FIG. 1 ), e.g., QR codes, barcodes, text, and any other scannable codes.
- the volume dimensioning system 100 , 100 a may analyze the 3D image data ( 128 , FIG. 1 ) and the holographic model 302 to detect and decode 3D encoded identifiers ( 702 ) on an exterior surface of the holographic model.
- 3D encoded identifiers 702 may include any three-dimensional feature on or of an exterior surface of the target object 120 detectable as a digital mark and decodable by the volume dimensioning system 100 , 100 a.
- 3D encoded identifiers 702 may include encoded three-dimensional features imperceptible or barely perceptible to the naked eye, or to the 2D imager ( 104 , FIG.
- 3D encoded identifiers 702 may be incorporated into blockchain-driven or other types of distributed ledgers configured for tracking the target object 120 through a supply chain.
- the 3D encoded identifier 702 may incorporate a unique identifier of the target object 120 , uniquely identifying its contents, origin, destination, handling instructions, perishability, and any other aspects and details specific to the target object, allowing the target object to be individually and fully traceable at any point within the supply chain.
- the full history of the target object 120 may be transparent to parties at any point of the supply chain or beyond it, e.g., manufacturers and wholesalers, third party transporters and logistics managers, retailers and trade associations, consumers and watchdog organizations, or public authorities.
- the volume dimensioning system 100 b may be implemented and may function similarly to the volume dimensioning systems 100 , 100 a of FIGS. 1 and 6 a , except that the volume dimensioning system 100 b may display the 2D video stream ( 126 , FIG. 2 ), the 3D image data ( 128 , FIG. 1 ) and the holographic model 302 via the AR viewing device 136 .
- the volume dimensioning system 100 b may be compatible with Microsoft HoloLens 1 ⁇ 2, MagicLeap, Oculus Rift, and other similar AR, VR, and MR wearable retinal display devices, broadcasting in real time or near real time to the AR viewing device 136 (via the wireless link 112 a, FIG.
- the AR viewing device 136 may be immediately proximate to, and physically linked to, the mobile device 102 , receiving the 2D image stream 126 and 3D image data 128 via the physical link.
- the AR viewing device 136 may detect and respond to control input provided by the wearer, adjusting the AR/MR display accordingly (either individually or in conjunction with the mobile device 102 capturing the 2D video stream 126 and 3D image data 128 .
- the wearer may move his/her hand 802 slowly downward ( 804 ) within the field of view of the AR viewing device 136 , which the volume dimensioning system 100 b may interpret as a command to rotate downward ( 806 ) the holographic model 302 (e.g., relative to its forward edge 306 a ) while the 2D image stream 126 featuring the target object 120 remains fixed in its real-world orientation.
- the wearer may similarly submit control input to move or resize the holographic model 302 relative to the target object 120 or access any encoded information detected and decoded via the 2D image stream 126 (e.g., QR codes, barcodes, and other 2D encoding) or the holographic model 302 (e.g., 3D encoded identifiers ( 702 , FIG. 7 )).
- any encoded information detected and decoded via the 2D image stream 126 e.g., QR codes, barcodes, and other 2D encoding
- the holographic model 302 e.g., 3D encoded identifiers ( 702 , FIG. 7 )
- the volume dimensioning system 100 b may recognize and respond to verbal control input ( 808 ) provided by the wearer and detected by the AR viewing device 136 .
- the wearer of the AR viewing device 136 may view and manipulate the holographic model 302 via the wireless link 112 a from a physical location remotely located from the target object 120 and the mobile device 102 .
- the volume dimensioning system 100 c may be implemented and may function similarly to the volume dimensioning systems 100 , 100 a - b of FIGS. 1, 6A, and 8A , except that the volume dimensioning system 100 c may compare the holographic model 302 to additional reference data (e.g., stored to the memory 134 , FIG. 1 ) in order to make additional determinations with respect to the target object ( 120 , FIG. 1 ) to which the holographic model corresponds after the dimensions 402 of the target object have been determined to a sufficient level of accuracy or confidence.
- additional reference data e.g., stored to the memory 134 , FIG. 1
- the volume dimensioning system 100 c may compare the dimensions 402 of the target object 120 to the dimensions of shipping boxes ( 902 ) or predetermined templates ( 904 ) corresponding to shipping boxes or other known objects having known dimensions (e.g., stored to memory 138 or accessible via cloud-based storage 132 or remote databases stored on physical servers 130 ).
- the volume dimensioning system 100 c may display for the user's selection (e.g., via a searchable menu 906 ) templates 904 corresponding to storage containers, storage bins, or storage locations and sublocations within racking, shelving or organizing systems of various sizes.
- the user may compare the determined dimensions 402 of the target object to a predetermined template 904 to determine, e.g., whether the target object 120 corresponds to a template 904 , whether the target object will fit inside a larger object or within a given shipping space, or for auditing or certifying a dimension measurement, or for calibrating/verifying the accuracy of the volume dimensioning system 100 c.
- the user may manually enter template dimensions to which the measured dimensions 402 of the target object 120 may be compared (e.g., if the orientations of a template 904 do not precisely match a target object 120 to which the template dimensions may otherwise correspond).
- the volume dimensioning system 100 c may infer this information from what is known about similarly sized shipping boxes 902 or their contents.
- the user may fill in the dimensions 402 of the target object 120 based on a corresponding template 904 that approximates or matches the dimensions of the target object. The user may create a new template by measuring a target object 120 and adding its dimensions 402 as a new known object.
- the mobile device 102 of the volume dimensioning system 100 c may be implemented by an operator at an aircraft gate to analyze carry-on items, generate holographic models 302 corresponding thereto, and compare the dimensions 402 of the corresponding carry-on items to standard-sized luggage items ( 908 ). In this way, the volume dimensioning system 100 c may make rapid-response determinations, e.g., whether a particular carry-on item or piece of luggage should be checked instead, or whether the said item or piece is likely to fit under a seat or within an overhead bin (e.g., corresponding to a particular aircraft type, cabin layout, seat class, seat type, or assigned seat).
- rapid-response determinations e.g., whether a particular carry-on item or piece of luggage should be checked instead, or whether the said item or piece is likely to fit under a seat or within an overhead bin (e.g., corresponding to a particular aircraft type, cabin layout, seat class, seat type, or assigned seat).
- the user may create a template called “SMALL BOX” ( 904 a ) having predefined dimensions of 8′′ ⁇ 11.5′′ ⁇ 6.75′′, measuring a target object corresponding to these dimensions via the volume dimensioning system 100 c to calibrate the volume dimensioning system and “learn” the new template for future reference.
- SMALL BOX 904 a
- the volume dimensioning system 100 c may identify one or more target objects 120 (e.g., within the FOV of the 2D imager 104 and 3D imager 106 ) by recognizing the object as a known object based on a comparison of 2D image data 126 and/or 3D spatial information 128 to records of other, similar objects (e.g., stored to memory 138 , cloud-based storage 130 , or remote physical servers 132 ).
- the volume dimensioning system 100 c may identify the target object 120 within the FOV of the mobile device 102 i by comparing the target object to reference examples of known 2D imagery or 3D image models (e.g., computer-aided design (CAD) models) to identify similarities in size, dimensions, shape features, or other aspects of the reference images to build confidence in the similarly of the target object to another known object. If sufficient confidence is achieved, the volume dimensioning system 100 c may positively identify the target object as equivalent to a known reference object; otherwise, additional scanning operations may be performed to reinforce confidence levels or confirm an identification. Reference comparison data may be used by the volume dimensioning system 100 c to supplement the holographic model 302 or aid in dimensioning, measurement and analysis operations.
- CAD computer-aided design
- the volume dimensioning system 100 c may compare the target object 120 to a particular suitcase 910 , noting any similarities in proportions and the presence or absence of, e.g., casters 912 , a handle 914 , or identification tagging 916 . If the observed similarities are strong enough, the volume dimensioning system 100 c may conclude with sufficient confidence that the target object 120 is equivalent to the suitcase 910 , or that the target object 120 is indeed a suitcase of that particular model or size.
- the volume dimensioning system 100 c may identify the target object 120 as, for example, a specific class of object, a subobject, or a component or part of a larger or more complex device based on comparison of the target object to reference data.
- the caster 912 may be removed from the suitcase 910 and scanned by the volume dimensioning system 100 c, which may compare the caster to similar parts or components (e.g., within parts catalogs accessible via memory 138 , cloud-based storage 130 , or remote physical servers 132 ) and determine specific object data corresponding to the caster, e.g., a manufacturer, make, or model number.
- pieces or parts may be removed from complex devices or systems, e.g., a pump removed from an engine, and scanned to identify the specific object (and acquire corresponding object data thereof) via comparison with reference manufacturer databases.
- the volume dimensioning system 100 d may be implemented and may function similarly to the volume dimensioning system 100 c of FIG. 9 , except that the target object of the volume dimensioning system 100 d may include non-cuboid objects or non-solid objects.
- the operator of the mobile device 102 may wish to determine a volume of concrete necessary to fill the driveway 1002 of the residence 1004 to a desired depth. The operator may position the mobile device 102 to capture a 2D image stream ( 126 , FIG. 1 ) and 3D image data ( 128 , FIG. 1 ) corresponding to the driveway 1002 , generating an appropriate holographic model 302 of the driveway and determining the dimensions 402 of the holographic model 302 .
- volume dimensioning system 100 d may be implemented to accurately measure the dimensions and volume of gaps, holes, or portions otherwise missing from an object, e.g., potholes in a street or road; open wounds; cracks or fissures in a wall or structure; and empty pools or reservoirs.
- a method 1100 for volume dimensioning of a substantially cuboid target object may be implemented by embodiments of the volume dimensioning systems 100 , 100 a - d.
- a two-dimensional (2D) camera or imaging system attached to a mobile device captures a 2D image stream corresponding to a field of view (FOV) and including at least one target object within the FOV.
- the 2D image stream may include multiple target objects.
- a three-dimensional (3D) imager of the mobile device collects 3D image data corresponding to the FOV; the 3D image data may include a cloud or set of points corresponding to each target object, where each point comprises a coordinate set relative to the FOV and distance information, e.g., a distance of the point from the mobile device
- a processor of the mobile device distinguishes the target object from the FOV (e.g., locks onto the target object from within the surrounding environment) by analyzing at least one of the 2D image stream and the 3D image data.
- the volume dimensioning system may analyze the 3D point cloud to identify planar surfaces, edges, or vertices of the target object, e.g., based on the alignment or relative distance of points or groups of points within the point cloud.
- the processor generates (and the display surface displays) a holographic model corresponding to each positively identified target object by correlating the 2D image stream and the corresponding point cloud.
- the holographic model includes surfaces and adjustable surface guides corresponding to the planar surfaces of the target object, edges and adjustable edge guides corresponding to the edges of the target object, and vertices and adjustable vertex guides corresponding to the vertices of the target object
- the processor determines one or more dimensions of the target object by measuring the holographic model. For example, the processor may measure sets of parallel edges to determine a given dimension to a particular confidence level.
- the processor detects object identifiers of the target object by analyzing the holographic model.
- the processor may detect 2D encoded information (e.g., barcodes, QR codes) or 3D encoded information integrated into a surface of the target object.
- the processor decodes the identified 2D and 3D object identifiers to obtain object data corresponding to the target object, e.g., supply chain or shipping data uniquely identifying the target object within the supply chain or chain of custody.
- the processor may supplement the object data with additional data, e.g., uploading the 2D image stream and 3D holographic model to a cloud archive, submitting the measured dimensions for auditing or further analysis, adding geolocation and timestamp data to assure point-in time condition and location continuity of the target object.
- embodiments of the methods disclosed herein may include one or more of the steps described herein. Further, such steps may be carried out in any desired order and two or more of the steps may be carried out simultaneously with one another. Two or more of the steps disclosed herein may be combined in a single step, and in some embodiments, one or more of the steps may be carried out as two or more sub-steps. Further, other steps or sub-steps may be carried in addition to, or as substitutes to one or more of the steps disclosed herein.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Computer Graphics (AREA)
- Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Electromagnetism (AREA)
- Toxicology (AREA)
- Artificial Intelligence (AREA)
- Computer Hardware Design (AREA)
- User Interface Of Digital Computer (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
- The present application is related to and claims the benefit of the earliest available effective filing dates from the following listed applications (the “Related Applications”) (e.g., claims earliest available priority dates for other than provisional patent applications (e.g., under 35 USC § 120 as a continuation in part) or claims benefits under 35 USC § 119(e) for provisional patent applications, for any and all parent, grandparent, great-grandparent, etc. applications of the Related Applications).
- U.S. patent application Ser. No. 16/390,562 entitled SYSTEM FOR VOLUME DIMENSIONING VIA HOLOGRAPHIC SENSOR FUSION, filed Apr. 22, 2019, issuing Feb. 11, 2020 as U.S. Pat. No. 10,559,086;
- U.S. patent application Ser. No. 15/156,149 entitled SYSTEM AND METHODS FOR VOLUME DIMENSIONING FOR SUPPLY CHAINS AND SHELF SETS, filed May 16, 2016, which issued Apr. 23, 2019 as U.S. Pat. No. 10,268,892;
- U.S. Provisional Patent Application Ser. No. 62/694,764 entitled SYSTEM FOR
VOLUME DIMENSIONING VIA 2D/3D SENSOR FUSION, filed Jul. 6, 2018; - and U.S. Provisional Patent Application Ser. No. 62/162,480 entitled SYSTEMS AND METHODS FOR COMPREHENSIVE SUPPLY CHAIN MANAGEMENT VIA MOBILE DEVICE, filed May 15, 2015.
- Said U.S. patent application Ser. Nos. 16/390,562; 15/156,149; 62/162,480; and 62/694,764 are herein incorporated by reference in their entirety.
- While many smartphones, pads, tablets, and other mobile computing devices are equipped with front-facing or rear-facing cameras, these devices may now be equipped with three-dimensional imaging systems incorporating cameras configured to detect infrared radiation combined with infrared or laser illuminators (e.g., light detection and ranging (LIDAR) systems) to enable the camera to derive depth information. It may be desirable for a mobile device to capture 3D images of objects, or two-dimensional images with depth information, and derive from the captured imagery additional information about the objects portrayed, such as the dimensions of the objects or other details otherwise accessible through visual comprehension, such as significant markings, encoded information, or visible damage.
- A wearable augmented reality (AR) device (e.g., virtual reality (VR), mixed reality (MR) device is disclosed. The wearable device communicates with a system for volume dimensioning via two-dimensional (2D)/three-dimensional (3D) sensor fusion. The AR device display to the user a two-dimensional (2D) image stream of a field of view (FOV) including target objects and 3D imaging data of the FOV including point clouds of each target object within the FOV. Target objects are identified by correlating the 2D and 3D image streams and generating a holographic model of the target object overlaid on the video stream, with adjustable surface, edge, and vertex guides. Precise dimensions and unique object identifiers of the target object are determined and decoded and displayed to the user via the wearable device; the user may manipulate or adjust the display via gestures or commands detectable by the wearable device.
- This Summary is provided solely as an introduction to subject matter that is fully described in the Detailed Description and Drawings. The Summary should not be considered to describe essential features nor be used to determine the scope of the Claims. Moreover, it is to be understood that both the foregoing Summary and the following Detailed Description are example and explanatory only and are not necessarily restrictive of the subject matter claimed.
- The detailed description is described with reference to the accompanying figures. The use of the same reference numbers in different instances in the description and the figures may indicate similar or identical items. Various embodiments or examples (“examples”) of the present disclosure are disclosed in the following detailed description and the accompanying drawings. The drawings are not necessarily to scale. In general, operations of disclosed processes may be performed in an arbitrary order, unless otherwise provided in the claims. In the drawings:
-
FIG. 1 is a block diagram illustrating a system for volume dimensioning via sensor fusion, in accordance with example embodiments of this disclosure; -
FIGS. 2A and 2B are diagrammatic illustrations of a mobile device of the volume dimensioning system ofFIG. 1 ; -
FIGS. 3A and 3B are diagrammatic illustrations of a mobile device of the volume dimensioning system ofFIG. 1 ; -
FIGS. 4A through 4C are diagrammatic illustrations of a mobile device of the volume dimensioning system ofFIG. 1 ; -
FIGS. 5A and 5B are diagrammatic illustrations of a mobile device of the volume dimensioning system ofFIG. 1 ; -
FIGS. 6A and 6B illustrate operations of, and a mobile device of, the volume dimensioning system ofFIG. 1 ; -
FIG. 7 is a diagrammatic illustration of a mobile device of the volume dimensioning system ofFIG. 1 ; -
FIGS. 8A and 8B are diagrammatic illustrations of an augmented reality (AR) device of the volume dimensioning system ofFIG. 1 ; -
FIGS. 9A and 9B illustrate operations of, and the mobile device of, the volume dimensioning system ofFIG. 1 ; -
FIGS. 10A and 10B illustrate operations of, and the mobile device of, the volume dimensioning system ofFIG. 1 ; - and
FIG. 11 is a flow diagram illustrating a method for volume dimensioning via sensor fusion, in accordance with example embodiments of this disclosure. - Before explaining one or more embodiments of the disclosure in detail, it is to be understood that the embodiments are not limited in their application to the details of construction and the arrangement of the components or steps or methodologies set forth in the following description or illustrated in the drawings. In the following detailed description of embodiments, numerous specific details may be set forth in order to provide a more thorough understanding of the disclosure. However, it will be apparent to one of ordinary skill in the art having the benefit of the instant disclosure that the embodiments disclosed herein may be practiced without some of these specific details. In other instances, well-known features may not be described in detail to avoid unnecessarily complicating the instant disclosure.
- As used herein a letter following a reference numeral is intended to reference an embodiment of the feature or element that may be similar, but not necessarily identical, to a previously described element or feature bearing the same reference numeral (e.g., 1, 1 a, 1 b). Such shorthand notations are used for purposes of convenience only and should not be construed to limit the disclosure in any way unless expressly stated to the contrary.
- Further, unless expressly stated to the contrary, “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by anyone of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
- In addition, use of “a” or “an” may be employed to describe elements and components of embodiments disclosed herein. This is done merely for convenience and “a” and “an” are intended to include “one” or “at least one,” and the singular also includes the plural unless it is obvious that it is meant otherwise.
- Finally, as used herein any reference to “one embodiment” or “some embodiments” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment disclosed herein. The appearances of the phrase “in some embodiments” in various places in the specification are not necessarily all referring to the same embodiment, and embodiments may include one or more of the features expressly described or inherently present herein, or any combination of sub-combination of two or more such features, along with any other features which may not necessarily be expressly described or inherently present in the instant disclosure.
- A system for volume dimensioning via sensor fusion is disclosed. In embodiments, the volume dimensioning system includes two-dimensional (2D) and three-dimensional (3D) image sensors incorporated into or attached to a mobile device, e.g., a smartphone, tablet, phablet, or like portable processor-enabled device. The 2D imager captures an image stream or sequence (e.g., streaming video at 30 or 60 fps) corresponding to a field of view of the device, while the 3D imager creates a point cloud or similar point set (e.g., a 3D mesh representation) of the field of view. The image sources can be fused to detect a target object within the field of view (e.g., a shipping container or like cuboid object traveling through a supply chain). By analyzing the fused imagery, the target object can be accurately measured and any encoded information relevant to the object decoded and updated
- Referring to
FIG. 1 , asystem 100 for volume dimensioning via sensor fusion is disclosed. Thevolume dimensioning system 100 may include amobile device 102 capable of being carried by an operator or user, themobile device 102 including 2D image sensors 104 (e.g., a visible-light camera),3D image sensors 106, image andcontrol processors 108, a touch-sensitive display surface 110, and awireless transceiver 112. Themobile device 102 may additionally include aclock 114 or time sensor, a Global Positioning System (GPS)receiver 116 or similar position sensor for determining a current position of the mobile device, and an inertial measurement unit 118 (IMU) or similar inertial sensor (e.g., accelerometer, magnetometer, gyrometer, compass) for determining a current orientation of the mobile device (or for tracking the orientation, and the rate of change thereof, over time). Instead of, or in addition to,onboard IMUs 118 of themobile device 102, thevolume dimensioning system 100 may incorporate IMUs integrated into the2D image sensors 104 or into the3D image sensors 106. The3D image sensors 106 may include imaging systems including infrared illuminators combined with multiple embedded cameras (e.g., Intel RealSense or other like triangulating systems), laser-based light detection and ranging (LIDAR) systems incorporating onboard photodetectors to track reflected beams, time of flight (ToF) camera systems, or any other like sensor system capable of producing 3D spatial information of proximate objects. As noted above, the3D image sensors 106 may incorporate inertial or orientation sensors or combinations thereof, e.g., accelerometers, gyroscopes, and compasses. - In embodiments, the
mobile device 102 may be oriented toward atarget object 120 in such a way that the2D image sensors 3D image sensors 106 simultaneously capture image data from a common field of view in which thetarget object 120 is situated. For example, thetarget object 120 may include a shipping box or container currently traveling through a supply chain, e.g., from a known origin to a known destination. Thetarget object 120 may be freestanding on a floor, table, or other flat surface; in some embodiments the target object may be secured to apallet 122 or similar structural foundation, either individually or in a group of such objects, for storage or transport. Thetarget object 120 is preferably substantially cuboid in shape, e.g., having six rectangular surfaces (120 a-c) intersecting at right angles. In embodiments, thetarget object 120 may not itself be perfectly cuboid but may fit perfectly within a minimumcuboid volume 124 of determinable dimensions (e.g., the minimum cuboid volume necessary to fully surround or encompass the target object). In embodiments, the precise dimensions of thetarget object 120 may be unknown to the operator of themobile device 102 but determinable via sensor fusion of therespective sensor inputs 2D imager 3D imager 106. - In embodiments, the
volume dimensioning system 100 may detect thetarget object 120 via a single sensor array, e.g., either the2D image sensors 104 or the3D image sensors 106, as opposed to a combination of 2D and 3D image sensors fusing their respective data outputs to aid detection. Similarly, thevolume dimensioning system 100 may simultaneously detect and analyze multiple target objects 120 within a particular field of view. - In embodiments, the
wireless transceiver 112 may enable the establishment of wireless links to remote sources, e.g.,physical servers 130 and cloud-basedstorage 132. For example, thewireless transceiver 112 may establish awireless link 112 a to aremote operator 134 situated at a physical distance from themobile device 102 and thetarget object 120, such that the remote operator may visually interact with thetarget object 120 and submit control input to themobile device 102. Similarly, thewireless transceiver 112 may establish awireless link 112 a to an augmented reality (AR) viewing device 136 (e.g., a virtual reality (VR) or mixed reality (MR) device worn on the head of a viewer, or proximate to the viewer's eyes, and capable of displaying to the viewer real-world objects and environments, synthetic objects and environments, or combinations thereof). For example, theAR viewing device 136 may allow the user to interact with thetarget object 120 and/or the mobile device 102 (e.g., submitting control input to manipulate the field of view, or a representation of the target object situated therein) via physical, ocular, or aural control input detected by the AR viewing device. - In embodiments, the
mobile device 102 may include amemory 138 or other like means of data storage accessible to the image andcontrol processors 108, the memory capable of storing reference data accessible to thevolume dimensioning system 100 to make additional determinations with respect to thetarget object 120. - In embodiments, the
mobile device 102 may include amicrophone 140 for receiving aural control input from the user/operator, e.g., verbal commands to thevolume dimensioning system 100. - Referring to
FIG. 2A , themobile device 102 a may be implemented and may function similarly to themobile device 102 ofFIG. 1 , except that themobile device 102 may be oriented toward thetarget object 120 in such a way that the 2D imager (104,FIG. 1 ) captures an image stream (126,FIG. 1 ) of a field of view including the target object (or multiple objects). In embodiments, thedisplay surface 110 may display theimage stream 126 to the operator of themobile device 102 in real time or near real time. The operator of themobile device 102 may position or orient the target object 120 (or the mobile device relative to the target object) to optimally position the target object within the field of view for image analysis. For example, thetarget object 120 may be positioned proximate to a floor surface (202) and/or wall (204) of optimal contrast to the target object (e.g., in color, in texture) or lit so as to optimize the contrast between the target object and the floor/wall. - In some embodiments, the operator may position and orient the
mobile device 102 relative to thetarget object 120 so that the entire upper surface (206) of the target object is clearly visible to the2D imager 104, and that a forward vertical edge (208) and forward top corner or vertex (210) faces the 2D imager. In embodiments, the volume dimensioning system (100,FIG. 1 ) may analyze thetarget object 120 based on cumulative views captured from multiple orientations around thetarget object 120. - In embodiments, the image and control processors (108,
FIG. 1 ) may analyze theimage stream 126 or component frames thereof to distinguish thetarget object 120 from its surroundings. For example, the image andcontrol processors 108 may attempt to identify patterns and shifts in color and texture that may be indicative of thetarget object 120 or its parameters (e.g., surfaces, edges, vertices). - Referring now to
FIG. 2B , themobile device 102 b may be implemented and may function similarly to themobile device 102 a ofFIG. 2A , except that themobile device 102 b may display via thedisplay surface 110, in addition to the image stream (126,FIG. 1 ) captured by the 2D imager (104,FIG. 1 ), 3D image data (128,FIG. 1 ) captured by the 3D imager (106,FIG. 1 ). - In embodiments, the field of view of the
3D imager 106 may substantially correspond to the field of view of theimage stream 126 captured by the2D imager 104. The3D image data 128 may include a stream of pixel sets, each pixel set substantially corresponding to a frame of the2D image stream 126. Accordingly, the pixel set may include a point cloud 212 (e.g., point map) substantially corresponding to thetarget object 120. Eachpoint 214 of thepoint cloud 212 may include a coordinate set (e.g., XY) locating the point relative to the field of view (e.g., to the frame, to the pixel set) as well as plane angle and depth data of the point, e.g., the distance of the point from themobile device 102 - While the
2D imager 104 attempts to define thetarget object 120 within its field of view based on texture, color, or lighting analysis of theimage stream 126, the3D imager 106 may analyze depth information about thetarget object 120 and its environment as shown within its field of view. For example, the3D imager 106 may identify thefloor 202 as a plane of gradually increasing depth that meets an intersecting plane (e.g., arear wall 204 or acandidate plane surface 216 corresponding to a surface (120 a-c,FIG. 1 ) of the target object 120). Based on the intersections of the candidate plane surfaces 216 of the target object 120 (e.g., with each other or with the floor 202), the3D imager 106 may identify candidate edges 218. Similarly, the intersection of three plane surfaces, or the intersection of two candidate edges 218, may indicate acandidate vertex 220. - In some embodiments, the
volume dimensioning system 100 may account for imperfect data sets, e.g., gaps or holes in the point cloud, via plane identification. For example, the volume dimensioning system may analyze 3D spatial information 128 (independently or fused with the 2D image stream 126) to infer the planes of thetarget object 120, e.g., on the basis of a sufficient number of identified points aligned in a plane or nearly enough aligned (e.g., within a predetermined range) to derive the existence of a plane. By utilizing plane identification based solely on 3Dspatial information 128 collected by the3D imager 106, thevolume dimensioning system 100 may identify thetarget object 120 and its component planes quickly enough, or to a sufficient level of confidence, that fusion of2D image data 126 may not be necessary for optimal performance. In some embodiments, the volume dimensioning system may similarly infer the edges or vertices of thetarget object 120 based on the placement or alignment of individual points. - Referring to
FIG. 3A , themobile device 102 c may be implemented and may function similarly to themobile device 102 b ofFIG. 2B , except that with respect to themobile device 102 c, the volume dimensioning system (100,FIG. 1 ) may correlate the 2D image stream (126,FIG. 1 ) and 3D point cloud (212,FIG. 2B ) to identify thetarget object 120 and generate aholographic model 302 corresponding to the target object. - By correlating color, form, and texture analysis of the
image stream 126 with depth information from the3D point cloud 212, thevolume dimensioning system 100 may identify candidate parameters of thetarget object 120 and digitally represent these candidate parameters in theholographic model 302. For example, thevolume dimensioning system 100 may tentatively identify candidate surfaces 304, candidate edges 306, andcandidate vertices 308 of theholographic model 302, which candidate surfaces, edges, and vertices correspond to the real-world parameters (e.g.,surfaces 120 a-c,FIG. 1 ) of thetarget object 120. Based on the tentative dimensions of these candidate surfaces 304, candidate edges 306, andcandidate vertices 308, theholographic model 302 may be displayed to the operator of themobile device 102 c (e.g., via the display surface 110) as a three-dimensional object. In embodiments, the manner or style in which theholographic model 302 is displayed may vary based on operator preference or control input, e.g., wireframe, semitransparent solid, monochrome, multicolor. - In embodiments, the
volume dimensioning system 100 will generate theholographic model 302 by overlaying the thetarget object 120 with edge lines 310 (e.g., edge guides), vertex points 312 (e.g., vertex guides), and/or surface guides 314 (e.g., solid, shaded, textured planes, or planes of varying opaqueness) as the volume dimensioning system locks onto thetarget object 120, indicating to the user/operator that the target object has been positively identified. Further, the 2D image of thetarget object 120 may be transposed onto or otherwise incorporated into theholographic model 302, such that the 3Dholographic model 302 appears substantially similar to the user/operator as would the real target object 120 (e.g., when viewed directly or via the 2D imager 104). - In embodiments, the
volume dimensioning system 100 may be trained via machine learning to recognize and lock onto atarget object 120, positively identifying the target object and distinguishing the target object from its surrounding environment (e.g., the field of view of the2D imager 3D imager 106 including the target object as well as other candidate objects, which may additionally be locked onto as target objects and dimensioned). For example, thevolume dimensioning system 100 may include a recognition engine trained on positive and negative images of a particular object specific to a desired use case. As the recognition engine has access to location and timing data corresponding to each image or image stream (e.g., determined by aclock 114/GPS receiver 116 or similar position sensors of the embodyingmobile device 102 a or collected from image metadata), the recognition engine may be trained to specific latitudes, longitudes, and locations, such that the performance of the recognition engine may be driven in part by the current location of themobile device 102 a, the current time of day, the current time of year, or some combination thereof. - Referring to
FIG. 3B , themobile device 102 d may be implemented and may function similarly to themobile device 102 c ofFIG. 3A , except that with respect to themobile device 102 d, once theholographic model 302 is generated by the volume dimensioning system (100,FIG. 1 ), the operator may manipulate theholographic model 302 as displayed by thedisplay surface 110. For example, by sliding his/her finger (316) across the touch-sensitive display surface 110, the operator may move (318) theholographic model 302 relative to the display surface (e.g., and relative to theimage stream 126 and target object (or multiple objects simultaneously in the field of view) 102 portrayed thereby) or rotate (320) the holographic model. Similarly, the candidate parameters of theholographic model 302, e.g., edge guides 310, vertex guides 312, and surface guides 314 (or their corresponding candidate surfaces 304, candidate edges 306, and candidate vertices 308) may be shifted, resized, or corrected as shown below. In embodiments, theholographic model 302 may be manipulated based onaural control input 322 submitted by the operator (e.g., via a microphone (140,FIG. 1 ) disposed within themobile device 102 d). For example, thevolume dimensioning system 100 may respond to verbal commands from the operator, e.g., to shift or rotate theholographic model 302, or to select and shift anedge guide 310,vertex guide 312, orsurface guide 314. - Referring to
FIGS. 4A through 4C , themobile device 102 e may be implemented and may function similarly to themobile device 102 d ofFIG. 3B , except that with respect to themobile device 102 e, when the target object (120,FIG. 3B ) has been identified (e.g., the 3D point cloud (212,FIG. 2B ) has been fused and correlated with the 2D image stream (126,FIG. 1 ) of the target object (120) and the candidate edges (306,FIG. 3B ) and other candidate parameters of theholographic model 302 have been identified, the volume dimensioning system (100,FIG. 1 ) may measure theholographic model 302 to determine the dimensions of thetarget object 120 to which theholographic model 302 corresponds. - In embodiments, the
holographic model 302 of thecuboid target object 120 may itself correspond to a cuboid digital representation. For example, in a preferred view of the holographic model 302 (as shown byFIGS. 4A-C ) atop surface 304 a, aforward edge 306 a, and a top forward corner or vertex 308 (where the top surface and forward edge intersect) may all be displayed via thedisplay surface 110. Further, nine of the twelve edges of theholographic model 302 may be displayed, or three edges for each linear dimension of the holographic model. - In embodiments, the
volume dimensioning system 100 may determine the precise dimensions (402) of the target object 120 (e.g., x-axis width, y-axis breadth, z-axis height, volume) by measuring edges of theholographic model 302 corresponding thereto. For example, with respect to the z-axis height (404) of thetarget object 120, thevolume dimensioning system 100 may measure thevertical edges 306 a-c of theholographic model 302. By measuring two or three parallelvertical edges 306 a-c rather than a single edge, thevolume dimensioning system 100 may account for general model or technology variations, errors, or holes (e.g., incompletions, gaps) in the3D point cloud 212 which may skew individual edge measurements (particularly if the hole coincides with a vertex 308 (e.g., an endpoint of the edge) of theholographic model 302. Thevolume dimensioning system 100 would apply the same methodology in measuring the other parallel edges (306 d-f, 306 g-i;FIGS. 4B and 4C ) to produce a line length of all threedimensions - In embodiments, the
volume dimensioning system 100 may adjust the measuring process (e.g., based on control input from the operator) for increased accuracy or speed. For example, the measurement of a given dimension may be based on multiple readings or pollings of the holographic model 302 (e.g., by generating multiple holographic models per second on a frame-by-frame basis and selecting “good” measurements to generate a result set (e.g., 10 measurement sets) for averaging). Alternatively or additionally, the three measurements of thevertical edges 306 a-c may be averaged to determine a given dimension. Similarly, if twoedges 306 a-b within a single frame measure within a predetermined threshold (e.g., 5 mm), the measurement may be counted as a “good” reading for purposes of inclusion within a result set. In some embodiments, the confirmation tolerance may be increased by requiring all threeedges 306 a-c to be within the threshold variance for inclusion in the result set. - Referring in particular to
FIGS. 4B and 4C , in embodiments the x-axis width 406, y-axis breadth 408, andvolume 410 of the holographic model 302 (and therefore the target object 120) may similarly be determined by analysis and measurement of the x-axis edges 306 d-f and y-axis edges 306 g-i of theholographic model 302. In some embodiments, thevolume dimensioning system 100 may proceed at a reduced confidence level if measurements cannot be established at full confidence. For example, the exterior surface of thetarget object 120 may be matte-finished, light-absorbing, or otherwise treated in such a way that the volume dimensioning system (particularly with respect to the 2D imager 104) may have difficulty accurately determining or measuring surfaces, edges, and vertices of theholographic model 302. Under reduced-confidence conditions, thevolume dimensioning system 100 may, for example, reduce the number of minimum confirmations required for an acceptable measure (e.g., from 3 to 2) or analyze additional frames per second (e.g., sacrificing operational speed for enhanced accuracy). The confidence condition level (412) may be displayed to the user and stored in the dataset corresponding to thetarget object 120. - Referring to
FIGS. 5A and 5B , themobile device 102 f may be implemented and may function similarly to themobile device 102 e ofFIGS. 4A through 4C , except that with respect to themobile device 102 f, the volume dimensioning system (100,FIG. 1 ) may account for anypallets 122 or other foundations to which thetarget object 120 is attached, determining the minimumpossible dimensions 124 of the target object (e.g., based on the minimum possible amount of shelf space the target object attached to the pallet would occupy in a vehicle, in a warehouse, or elsewhere in the supply chain. Referring in particular toFIG. 5B , theholographic model 302 a may be implemented and may function similarly to theholographic model 302 ofFIGS. 4A-B , except that theholographic model 302 a may account for thepallet 122 in determining the minimum possible edges (502), minimum possible surfaces (504), and minimum possible volume of thetarget object 120 attached to thepallet 122. - Referring to
FIG. 6A , thevolume dimensioning system 100 a may be implemented and may function similarly to thevolume dimensioning system 100 ofFIG. 1 , except that thevolume dimensioning system 100 a may monitor the onboard IMU (118,FIG. 1 ) of the mobile device 102 (e.g., or inertial/orientation sensors integrated into the 2D imager (104,FIG. 1 ) or 3D imager (106,FIG. 1 ) to detect difficulty in the identification of candidate surfaces (304,FIGS. 3A-B ), candidate edges (306,FIGS. 3A-B ), and candidate vertices (306,FIGS. 3A-B ) from the holographic model 302 (FIGS. 3A-B ) corresponding to thetarget object 120. - For example, the
IMU 118 may detect excessive shifts in the orientation of themobile device 102 as the user (602) moves the mobile device around (604) and thevolume dimensioning system 100 a attempts to lock into the parameters of the target object via the2D image stream 126 and the3D image data 128. Similarly, theIMU 118 may notice rotational movement by theuser 604 around thetarget object 120 and take this movement into account in the generation of the 3Dholographic model 302. - Referring in particular to
FIG. 6B , themobile device 102 g may be implemented and may function similarly to themobile device 102 ofFIG. 6A , except that with respect to themobile device 102 g, thevolume dimensioning system 100 a may respond to the detection of certain types of motion by theIMU 118 by displaying surface guides 314 (e.g., planar guides), edge guides 310, and vertex guides 312 to assist theuser 602 in manually adjusting the candidate planes 310,edges 306 andcandidate vertices 308 of theholographic model 302. For example, theuser 602 may manipulate the edge guides 310, surface guides 314, and/or vertex guides 312 (e.g., via the touch-sensitive display surface 110) to precisely set the candidate edges 306 andcandidate vertices 308 so that thevolume dimensioning system 100 a may accurately measure theholographic model 302. In some embodiments, thevolume dimensioning system 100 a may sense a user/operator manipulating asurface guide 314,edge guide 310, orvertex guide 312 and understand intuitively where the user is intending to place the vertex guide (312 a), in which case thevertex guide 312 would automatically snap into the nearest logical place. - Referring now to
FIG. 7 , themobile device 102 h may be implemented and may function similarly to themobile device 102 g ofFIG. 6B and themobile devices FIGS. 1 through 5B , except that with respect to themobile device 102 h, the volume dimensioning system (100,FIG. 1 ; 100 a,FIG. 6A ) may analyze the 2D image stream (126,FIG. 1 ) to detect and decode any encoded information on an exterior surface of the target object (120,FIG. 1 ), e.g., QR codes, barcodes, text, and any other scannable codes. Similarly, thevolume dimensioning system FIG. 1 ) and theholographic model 302 to detect and decode 3D encoded identifiers (702) on an exterior surface of the holographic model. For example, 3D encodedidentifiers 702 may include any three-dimensional feature on or of an exterior surface of thetarget object 120 detectable as a digital mark and decodable by thevolume dimensioning system identifiers 702 may include encoded three-dimensional features imperceptible or barely perceptible to the naked eye, or to the 2D imager (104,FIG. 1 ), e.g., watermarks, grooves, ridges, raised or indented patterns, geometric patterns, textures, or any combination thereof. In embodiments, 3D encodedidentifiers 702 may be incorporated into blockchain-driven or other types of distributed ledgers configured for tracking thetarget object 120 through a supply chain. For example, the 3D encodedidentifier 702 may incorporate a unique identifier of thetarget object 120, uniquely identifying its contents, origin, destination, handling instructions, perishability, and any other aspects and details specific to the target object, allowing the target object to be individually and fully traceable at any point within the supply chain. Accordingly, the full history of thetarget object 120 may be transparent to parties at any point of the supply chain or beyond it, e.g., manufacturers and wholesalers, third party transporters and logistics managers, retailers and trade associations, consumers and watchdog organizations, or public authorities. - Referring to
FIGS. 8A and 8B , thevolume dimensioning system 100 b may be implemented and may function similarly to thevolume dimensioning systems FIGS. 1 and 6 a, except that thevolume dimensioning system 100 b may display the 2D video stream (126,FIG. 2 ), the 3D image data (128,FIG. 1 ) and theholographic model 302 via theAR viewing device 136. For example, thevolume dimensioning system 100 b may be compatible with Microsoft HoloLens ½, MagicLeap, Oculus Rift, and other similar AR, VR, and MR wearable retinal display devices, broadcasting in real time or near real time to the AR viewing device 136 (via thewireless link 112 a,FIG. 1 ) as an alternative to, or in addition to, the mobile device 102 (FIG. 1 ). In embodiments, theAR viewing device 136 may be immediately proximate to, and physically linked to, themobile device 102, receiving the2D image stream 3D image data 128 via the physical link. - In embodiments, the
AR viewing device 136 may detect and respond to control input provided by the wearer, adjusting the AR/MR display accordingly (either individually or in conjunction with themobile device 102 capturing the2D video stream 3D image data 128. For example, the wearer may move his/herhand 802 slowly downward (804) within the field of view of theAR viewing device 136, which thevolume dimensioning system 100 b may interpret as a command to rotate downward (806) the holographic model 302 (e.g., relative to itsforward edge 306 a) while the2D image stream 126 featuring thetarget object 120 remains fixed in its real-world orientation. The wearer may similarly submit control input to move or resize theholographic model 302 relative to thetarget object 120 or access any encoded information detected and decoded via the 2D image stream 126 (e.g., QR codes, barcodes, and other 2D encoding) or the holographic model 302 (e.g., 3D encoded identifiers (702,FIG. 7 )). - In embodiments, the
volume dimensioning system 100 b may recognize and respond to verbal control input (808) provided by the wearer and detected by theAR viewing device 136. In some embodiments, the wearer of theAR viewing device 136 may view and manipulate theholographic model 302 via thewireless link 112 a from a physical location remotely located from thetarget object 120 and themobile device 102. - Referring to
FIG. 9A , thevolume dimensioning system 100 c may be implemented and may function similarly to thevolume dimensioning systems FIGS. 1, 6A, and 8A , except that thevolume dimensioning system 100 c may compare theholographic model 302 to additional reference data (e.g., stored to thememory 134,FIG. 1 ) in order to make additional determinations with respect to the target object (120,FIG. 1 ) to which the holographic model corresponds after thedimensions 402 of the target object have been determined to a sufficient level of accuracy or confidence. - For example, the
volume dimensioning system 100 c may compare thedimensions 402 of thetarget object 120 to the dimensions of shipping boxes (902) or predetermined templates (904) corresponding to shipping boxes or other known objects having known dimensions (e.g., stored tomemory 138 or accessible via cloud-basedstorage 132 or remote databases stored on physical servers 130). Thevolume dimensioning system 100 c may display for the user's selection (e.g., via a searchable menu 906) templates 904 corresponding to storage containers, storage bins, or storage locations and sublocations within racking, shelving or organizing systems of various sizes. The user may compare thedetermined dimensions 402 of the target object to a predetermined template 904 to determine, e.g., whether thetarget object 120 corresponds to a template 904, whether the target object will fit inside a larger object or within a given shipping space, or for auditing or certifying a dimension measurement, or for calibrating/verifying the accuracy of thevolume dimensioning system 100 c. Alternatively, the user may manually enter template dimensions to which the measureddimensions 402 of thetarget object 120 may be compared (e.g., if the orientations of a template 904 do not precisely match atarget object 120 to which the template dimensions may otherwise correspond). If the object data (e.g., as decoded from 2D encoding or 3D encoded identifiers (702,FIG. 7 )) corresponding to thetarget object 120 does not provide sufficient information about the target object (e.g., guidelines for storage, transport, or perishability), thevolume dimensioning system 100 c may infer this information from what is known about similarlysized shipping boxes 902 or their contents. Similarly, the user may fill in thedimensions 402 of thetarget object 120 based on a corresponding template 904 that approximates or matches the dimensions of the target object. The user may create a new template by measuring atarget object 120 and adding itsdimensions 402 as a new known object. - Referring to
FIG. 9B , themobile device 102 of thevolume dimensioning system 100 c may be implemented by an operator at an aircraft gate to analyze carry-on items, generateholographic models 302 corresponding thereto, and compare thedimensions 402 of the corresponding carry-on items to standard-sized luggage items (908). In this way, thevolume dimensioning system 100 c may make rapid-response determinations, e.g., whether a particular carry-on item or piece of luggage should be checked instead, or whether the said item or piece is likely to fit under a seat or within an overhead bin (e.g., corresponding to a particular aircraft type, cabin layout, seat class, seat type, or assigned seat). For example, the user may create a template called “SMALL BOX” (904 a) having predefined dimensions of 8″×11.5″×6.75″, measuring a target object corresponding to these dimensions via thevolume dimensioning system 100 c to calibrate the volume dimensioning system and “learn” the new template for future reference. - In some embodiments, the
volume dimensioning system 100 c may identify one or more target objects 120 (e.g., within the FOV of the2D imager 2D image data 126 and/or 3Dspatial information 128 to records of other, similar objects (e.g., stored tomemory 138, cloud-basedstorage 130, or remote physical servers 132). For example, thevolume dimensioning system 100 c may identify thetarget object 120 within the FOV of the mobile device 102 i by comparing the target object to reference examples of known 2D imagery or 3D image models (e.g., computer-aided design (CAD) models) to identify similarities in size, dimensions, shape features, or other aspects of the reference images to build confidence in the similarly of the target object to another known object. If sufficient confidence is achieved, thevolume dimensioning system 100 c may positively identify the target object as equivalent to a known reference object; otherwise, additional scanning operations may be performed to reinforce confidence levels or confirm an identification. Reference comparison data may be used by thevolume dimensioning system 100 c to supplement theholographic model 302 or aid in dimensioning, measurement and analysis operations. - For example, the
volume dimensioning system 100 c may compare thetarget object 120 to aparticular suitcase 910, noting any similarities in proportions and the presence or absence of, e.g.,casters 912, ahandle 914, or identification tagging 916. If the observed similarities are strong enough, thevolume dimensioning system 100 c may conclude with sufficient confidence that thetarget object 120 is equivalent to thesuitcase 910, or that thetarget object 120 is indeed a suitcase of that particular model or size. - In some embodiments, the
volume dimensioning system 100 c may identify thetarget object 120 as, for example, a specific class of object, a subobject, or a component or part of a larger or more complex device based on comparison of the target object to reference data. For example, thecaster 912 may be removed from thesuitcase 910 and scanned by thevolume dimensioning system 100 c, which may compare the caster to similar parts or components (e.g., within parts catalogs accessible viamemory 138, cloud-basedstorage 130, or remote physical servers 132) and determine specific object data corresponding to the caster, e.g., a manufacturer, make, or model number. Similarly, pieces or parts may be removed from complex devices or systems, e.g., a pump removed from an engine, and scanned to identify the specific object (and acquire corresponding object data thereof) via comparison with reference manufacturer databases. - Referring to
FIGS. 10A and 10B , thevolume dimensioning system 100 d may be implemented and may function similarly to thevolume dimensioning system 100 c ofFIG. 9 , except that the target object of thevolume dimensioning system 100 d may include non-cuboid objects or non-solid objects. For example, the operator of themobile device 102 may wish to determine a volume of concrete necessary to fill thedriveway 1002 of theresidence 1004 to a desired depth. The operator may position themobile device 102 to capture a 2D image stream (126,FIG. 1 ) and 3D image data (128,FIG. 1 ) corresponding to thedriveway 1002, generating an appropriateholographic model 302 of the driveway and determining thedimensions 402 of theholographic model 302. Similarly, thevolume dimensioning system 100 d may be implemented to accurately measure the dimensions and volume of gaps, holes, or portions otherwise missing from an object, e.g., potholes in a street or road; open wounds; cracks or fissures in a wall or structure; and empty pools or reservoirs. - Referring to
FIG. 11 , amethod 1100 for volume dimensioning of a substantially cuboid target object (e.g., having a height, a width, and a breadth) via sensor fusion may be implemented by embodiments of thevolume dimensioning systems - At a
step 1102, a two-dimensional (2D) camera or imaging system attached to a mobile device captures a 2D image stream corresponding to a field of view (FOV) and including at least one target object within the FOV. In some embodiments, the 2D image stream may include multiple target objects. - At a
step 1104, a three-dimensional (3D) imager of the mobile device collects 3D image data corresponding to the FOV; the 3D image data may include a cloud or set of points corresponding to each target object, where each point comprises a coordinate set relative to the FOV and distance information, e.g., a distance of the point from the mobile device - At a
step 1106, a processor of the mobile device distinguishes the target object from the FOV (e.g., locks onto the target object from within the surrounding environment) by analyzing at least one of the 2D image stream and the 3D image data. For example, the volume dimensioning system may analyze the 3D point cloud to identify planar surfaces, edges, or vertices of the target object, e.g., based on the alignment or relative distance of points or groups of points within the point cloud. - At a
step 1108, the processor generates (and the display surface displays) a holographic model corresponding to each positively identified target object by correlating the 2D image stream and the corresponding point cloud. The holographic model includes surfaces and adjustable surface guides corresponding to the planar surfaces of the target object, edges and adjustable edge guides corresponding to the edges of the target object, and vertices and adjustable vertex guides corresponding to the vertices of the target object - At a
step 1110, the processor determines one or more dimensions of the target object by measuring the holographic model. For example, the processor may measure sets of parallel edges to determine a given dimension to a particular confidence level. - At a
step 1112, the processor detects object identifiers of the target object by analyzing the holographic model. For example, the processor may detect 2D encoded information (e.g., barcodes, QR codes) or 3D encoded information integrated into a surface of the target object. - At a
step 1114, the processor decodes the identified 2D and 3D object identifiers to obtain object data corresponding to the target object, e.g., supply chain or shipping data uniquely identifying the target object within the supply chain or chain of custody. The processor may supplement the object data with additional data, e.g., uploading the 2D image stream and 3D holographic model to a cloud archive, submitting the measured dimensions for auditing or further analysis, adding geolocation and timestamp data to assure point-in time condition and location continuity of the target object. - It is to be understood that embodiments of the methods disclosed herein may include one or more of the steps described herein. Further, such steps may be carried out in any desired order and two or more of the steps may be carried out simultaneously with one another. Two or more of the steps disclosed herein may be combined in a single step, and in some embodiments, one or more of the steps may be carried out as two or more sub-steps. Further, other steps or sub-steps may be carried in addition to, or as substitutes to one or more of the steps disclosed herein.
- Although inventive concepts have been described with reference to the embodiments illustrated in the attached drawing figures, equivalents may be employed and substitutions made herein without departing from the scope of the claims. Components illustrated and described herein are merely examples of a system/device and components that may be used to implement embodiments of the inventive concepts and may be replaced with other devices and components without departing from the scope of the claims. Furthermore, any dimensions, degrees, and/or numerical ranges provided herein are to be understood as non-limiting examples unless otherwise specified in the claims.
Claims (8)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/786,268 US20200184667A1 (en) | 2015-05-15 | 2020-02-10 | System for Volume Dimensioning Via Holographic Sensor Fusion |
US17/114,066 US11669988B1 (en) | 2020-02-10 | 2020-12-07 | System and method for three-dimensional box segmentation and measurement |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201562162480P | 2015-05-15 | 2015-05-15 | |
US15/156,149 US10268892B1 (en) | 2015-05-15 | 2016-05-16 | System and methods for volume dimensioning for supply chains and shelf sets |
US201862694764P | 2018-07-06 | 2018-07-06 | |
US16/390,562 US10559086B1 (en) | 2015-05-15 | 2019-04-22 | System for volume dimensioning via holographic sensor fusion |
US16/786,268 US20200184667A1 (en) | 2015-05-15 | 2020-02-10 | System for Volume Dimensioning Via Holographic Sensor Fusion |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/390,562 Continuation US10559086B1 (en) | 2015-05-15 | 2019-04-22 | System for volume dimensioning via holographic sensor fusion |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/114,066 Continuation-In-Part US11669988B1 (en) | 2020-02-10 | 2020-12-07 | System and method for three-dimensional box segmentation and measurement |
Publications (1)
Publication Number | Publication Date |
---|---|
US20200184667A1 true US20200184667A1 (en) | 2020-06-11 |
Family
ID=69410665
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/390,562 Active US10559086B1 (en) | 2015-05-15 | 2019-04-22 | System for volume dimensioning via holographic sensor fusion |
US16/786,268 Abandoned US20200184667A1 (en) | 2015-05-15 | 2020-02-10 | System for Volume Dimensioning Via Holographic Sensor Fusion |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/390,562 Active US10559086B1 (en) | 2015-05-15 | 2019-04-22 | System for volume dimensioning via holographic sensor fusion |
Country Status (1)
Country | Link |
---|---|
US (2) | US10559086B1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2023220071A3 (en) * | 2022-05-10 | 2024-01-11 | Apple Inc. | Systems, methods, and graphical user interfaces for scanning and modeling environments |
US11941764B2 (en) | 2021-04-18 | 2024-03-26 | Apple Inc. | Systems, methods, and graphical user interfaces for adding effects in augmented reality environments |
US12020380B2 (en) | 2019-09-27 | 2024-06-25 | Apple Inc. | Systems, methods, and graphical user interfaces for modeling, measuring, and drawing using augmented reality |
EP4390641A1 (en) * | 2022-12-22 | 2024-06-26 | Mitsubishi Logisnext Co., Ltd. | Display device, cargo unloading support system, industrial vehicle, control method, and program |
WO2024191063A1 (en) * | 2023-03-10 | 2024-09-19 | 삼성전자 주식회사 | Method for providing user interface, and wearable electronic device supporting same |
US12131417B1 (en) | 2023-11-08 | 2024-10-29 | Apple Inc. | Devices, methods, and graphical user interfaces for depth-based annotation |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10559086B1 (en) * | 2015-05-15 | 2020-02-11 | 4DMobile, LLC | System for volume dimensioning via holographic sensor fusion |
WO2017200527A1 (en) * | 2016-05-16 | 2017-11-23 | Hewlett-Packard Development Company, L.P. | Generating a shape profile for a 3d object |
KR20210030147A (en) * | 2019-09-09 | 2021-03-17 | 삼성전자주식회사 | 3d rendering method and 3d rendering apparatus |
US11669988B1 (en) | 2020-02-10 | 2023-06-06 | 4DMobile, LLC | System and method for three-dimensional box segmentation and measurement |
CN113237423B (en) * | 2021-04-16 | 2023-09-05 | 北京京东乾石科技有限公司 | Article volume measuring device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130083173A1 (en) * | 2011-09-30 | 2013-04-04 | Kevin A. Geisner | Virtual spectator experience with a personal audio/visual apparatus |
US20140333666A1 (en) * | 2013-05-13 | 2014-11-13 | Adam G. Poulos | Interactions of virtual objects with surfaces |
US9344612B2 (en) * | 2006-02-15 | 2016-05-17 | Kenneth Ira Ritchey | Non-interference field-of-view support apparatus for a panoramic facial sensor |
US10268892B1 (en) * | 2015-05-15 | 2019-04-23 | 4Dmobile Llc | System and methods for volume dimensioning for supply chains and shelf sets |
US10559086B1 (en) * | 2015-05-15 | 2020-02-11 | 4DMobile, LLC | System for volume dimensioning via holographic sensor fusion |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5117098A (en) * | 1989-09-29 | 1992-05-26 | Symbol Technologies, Inc. | Multi-position trigger for control over aiming and symbol reading in a portable laser diode scanning head |
US5329106A (en) * | 1992-10-02 | 1994-07-12 | Psc, Inc. | Handle-less bar code scanning system |
US6947571B1 (en) | 1999-05-19 | 2005-09-20 | Digimarc Corporation | Cell phones with optical capabilities, and related applications |
US20020014533A1 (en) | 1995-12-18 | 2002-02-07 | Xiaxun Zhu | Automated object dimensioning system employing contour tracing, vertice detection, and forner point detection and reduction methods on 2-d range data maps |
US7128270B2 (en) | 1999-09-17 | 2006-10-31 | Silverbrook Research Pty Ltd | Scanning device for coded data |
US8239229B1 (en) | 2000-02-24 | 2012-08-07 | Richard Paiz | Parallel computer network and method for real time financial resource management, inventory control, and online purchasing |
US6908797B2 (en) | 2002-07-09 | 2005-06-21 | Semiconductor Energy Laboratory Co., Ltd. | Method of manufacturing a semiconductor device |
US8471812B2 (en) | 2005-09-23 | 2013-06-25 | Jesse C. Bunch | Pointing and identification device |
US8385971B2 (en) | 2008-08-19 | 2013-02-26 | Digimarc Corporation | Methods and systems for content processing |
CN102110332B (en) | 2009-12-24 | 2013-01-09 | 上海阿艾依智控系统有限公司 | Book registering and managing device based on computer vision and radio frequency identification technology |
US8775710B1 (en) * | 2011-02-07 | 2014-07-08 | Mobiledemand Lc | Mobile information handling apparatus |
US9332243B2 (en) | 2012-10-17 | 2016-05-03 | DotProduct LLC | Handheld portable optical scanner and method of using |
US10674135B2 (en) | 2012-10-17 | 2020-06-02 | DotProduct LLC | Handheld portable optical scanner and method of using |
US9464885B2 (en) | 2013-08-30 | 2016-10-11 | Hand Held Products, Inc. | System and method for package dimensioning |
-
2019
- 2019-04-22 US US16/390,562 patent/US10559086B1/en active Active
-
2020
- 2020-02-10 US US16/786,268 patent/US20200184667A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9344612B2 (en) * | 2006-02-15 | 2016-05-17 | Kenneth Ira Ritchey | Non-interference field-of-view support apparatus for a panoramic facial sensor |
US20130083173A1 (en) * | 2011-09-30 | 2013-04-04 | Kevin A. Geisner | Virtual spectator experience with a personal audio/visual apparatus |
US20140333666A1 (en) * | 2013-05-13 | 2014-11-13 | Adam G. Poulos | Interactions of virtual objects with surfaces |
US10268892B1 (en) * | 2015-05-15 | 2019-04-23 | 4Dmobile Llc | System and methods for volume dimensioning for supply chains and shelf sets |
US10559086B1 (en) * | 2015-05-15 | 2020-02-11 | 4DMobile, LLC | System for volume dimensioning via holographic sensor fusion |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US12020380B2 (en) | 2019-09-27 | 2024-06-25 | Apple Inc. | Systems, methods, and graphical user interfaces for modeling, measuring, and drawing using augmented reality |
US11941764B2 (en) | 2021-04-18 | 2024-03-26 | Apple Inc. | Systems, methods, and graphical user interfaces for adding effects in augmented reality environments |
WO2023220071A3 (en) * | 2022-05-10 | 2024-01-11 | Apple Inc. | Systems, methods, and graphical user interfaces for scanning and modeling environments |
EP4390641A1 (en) * | 2022-12-22 | 2024-06-26 | Mitsubishi Logisnext Co., Ltd. | Display device, cargo unloading support system, industrial vehicle, control method, and program |
WO2024191063A1 (en) * | 2023-03-10 | 2024-09-19 | 삼성전자 주식회사 | Method for providing user interface, and wearable electronic device supporting same |
US12131417B1 (en) | 2023-11-08 | 2024-10-29 | Apple Inc. | Devices, methods, and graphical user interfaces for depth-based annotation |
Also Published As
Publication number | Publication date |
---|---|
US10559086B1 (en) | 2020-02-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20200184667A1 (en) | System for Volume Dimensioning Via Holographic Sensor Fusion | |
US10268892B1 (en) | System and methods for volume dimensioning for supply chains and shelf sets | |
US11189044B2 (en) | Method and device for detecting object stacking state and intelligent shelf | |
Zollmann et al. | Augmented reality for construction site monitoring and documentation | |
US20200151450A1 (en) | Maintenance work support system and maintenance work support method | |
US9098905B2 (en) | System and method for determining position of a device | |
CN107402000A (en) | For the system and method relative to measuring instrument with reference to display device | |
US8831337B2 (en) | Method, system and computer program product for identifying locations of detected objects | |
WO2020076396A1 (en) | Real-world anchor in a virtual-reality environment | |
US20140267776A1 (en) | Tracking system using image recognition | |
EP3695381B1 (en) | Floor detection in virtual and augmented reality devices using stereo images | |
US10127667B2 (en) | Image-based object location system and process | |
US11380011B2 (en) | Marker-based positioning of simulated reality | |
US9529803B2 (en) | Image modification | |
CN106647738A (en) | Method and system for determining docking path of automated guided vehicle, and automated guided vehicle | |
US11263818B2 (en) | Augmented reality system using visual object recognition and stored geometry to create and render virtual objects | |
KR102075844B1 (en) | Localization system merging results of multi-modal sensor based positioning and method thereof | |
US10432915B2 (en) | Systems, methods, and devices for generating three-dimensional models | |
US20190287257A1 (en) | Method and system for measuring the distance to remote objects | |
US20170132561A1 (en) | Product delivery unloading assistance systems and methods | |
Li et al. | Using Kinect for monitoring warehouse order picking operations | |
WO2021207205A1 (en) | Geo-augmented field excursion for geological sites | |
Del Pizzo et al. | Reliable vessel attitude estimation by wide angle camera | |
CN113228117B (en) | Authoring apparatus, authoring method, and recording medium having an authoring program recorded thereon | |
US11669988B1 (en) | System and method for three-dimensional box segmentation and measurement |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: 4DMOBILE, LLC, IOWA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MILLER, MATTHEW D.;KUNERT, STEVEN R.;MCDERMOTT, CRAIG;AND OTHERS;REEL/FRAME:051771/0353 Effective date: 20190422 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |