GB2533676A - System and method for picking validation - Google Patents

System and method for picking validation Download PDF

Info

Publication number
GB2533676A
GB2533676A GB1517842.9A GB201517842A GB2533676A GB 2533676 A GB2533676 A GB 2533676A GB 201517842 A GB201517842 A GB 201517842A GB 2533676 A GB2533676 A GB 2533676A
Authority
GB
United Kingdom
Prior art keywords
patent application
physical attributes
item
application publication
sensed
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB1517842.9A
Other versions
GB2533676B (en
GB201517842D0 (en
Inventor
Chamberlain James
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hand Held Products Inc
Original Assignee
Hand Held Products Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US14/865,797 external-priority patent/US10810715B2/en
Application filed by Hand Held Products Inc filed Critical Hand Held Products Inc
Publication of GB201517842D0 publication Critical patent/GB201517842D0/en
Publication of GB2533676A publication Critical patent/GB2533676A/en
Application granted granted Critical
Publication of GB2533676B publication Critical patent/GB2533676B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/08Logistics, e.g. warehousing, loading or distribution; Inventory or stock management
    • G06Q10/087Inventory or stock management, e.g. order filling, procurement or balancing against orders
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B65CONVEYING; PACKING; STORING; HANDLING THIN OR FILAMENTARY MATERIAL
    • B65GTRANSPORT OR STORAGE DEVICES, e.g. CONVEYORS FOR LOADING OR TIPPING, SHOP CONVEYOR SYSTEMS OR PNEUMATIC TUBE CONVEYORS
    • B65G1/00Storing articles, individually or in orderly arrangement, in warehouses or magazines
    • B65G1/02Storage devices
    • B65G1/04Storage devices mechanical
    • B65G1/137Storage devices mechanical with arrangements or automatic control means for selecting which articles are to be removed
    • B65G1/1373Storage devices mechanical with arrangements or automatic control means for selecting which articles are to be removed for fulfilling orders in warehouses
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B65CONVEYING; PACKING; STORING; HANDLING THIN OR FILAMENTARY MATERIAL
    • B65GTRANSPORT OR STORAGE DEVICES, e.g. CONVEYORS FOR LOADING OR TIPPING, SHOP CONVEYOR SYSTEMS OR PNEUMATIC TUBE CONVEYORS
    • B65G1/00Storing articles, individually or in orderly arrangement, in warehouses or magazines
    • B65G1/02Storage devices
    • B65G1/04Storage devices mechanical
    • B65G1/137Storage devices mechanical with arrangements or automatic control means for selecting which articles are to be removed
    • B65G1/1373Storage devices mechanical with arrangements or automatic control means for selecting which articles are to be removed for fulfilling orders in warehouses
    • B65G1/1375Storage devices mechanical with arrangements or automatic control means for selecting which articles are to be removed for fulfilling orders in warehouses the orders being assembled on a commissioning stacker-crane or truck
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/08Logistics, e.g. warehousing, loading or distribution; Inventory or stock management
    • G06Q10/083Shipping

Landscapes

  • Business, Economics & Management (AREA)
  • Engineering & Computer Science (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Quality & Reliability (AREA)
  • Theoretical Computer Science (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Human Resources & Organizations (AREA)
  • General Physics & Mathematics (AREA)
  • Operations Research (AREA)
  • Development Economics (AREA)
  • Strategic Management (AREA)
  • Tourism & Hospitality (AREA)
  • Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • Accounting & Taxation (AREA)
  • Finance (AREA)
  • Mechanical Engineering (AREA)

Abstract

Verification of a picked item for shipping is achieved by sensing a physical characteristic of the item and comparing the detected characteristic with an expected characteristic. If the sensed and expected characteristics do not match an error alert is output to a user interface. The sensor may be a scale to measure weight, a camera using optical recognition or a measurement means to detect dimensions, such as an infra-red point cloud and associated receiver. A plurality of sensors may provide for more accurate verification. The sensor may be coupled to a mobile computer carried by a user, or may be incorporated into a powered industrial vehicle. The user may retrieve the expected characteristics by inputting a serial number, part number, location or the like e.g. via a keyboard or scanning a barcode. The alerts may be audio, such as a voice, visual such as text or graphics on a screen or a flashing light, haptic, or any combination. The expected characteristics may be pre-installed at set-up or machine learning may be employed, updating the expected characteristics from the sensed characteristics.

Description

SYSTEM AND METHOD FOR PICKING VALIDATION
CROSS-REFERENCE TO RELATED APPLICATION
[0001] This application claims the benefit of U.S. Patent Application Ser. No. 62/062,175 for Systems and Methods for Dimensioning, (filed October 10, 2014), which is hereby incorporated by reference in its entirety.
FIELD OF THE INVENTION
[0002] The present invention relates to computer-assisted picking and more specifically to systems/methods to validate a picking operation using a picked item's physical characteristics.
BACKGROUND
[0003] Order preparation (i.e., picking) is one operation in logistics. Picking consists of gathering a specified quantity of items from a location (e.g., a bin, a shelf in a warehouse, etc.) for some purpose (e.g., shipment, order fulfillment, assembly, etc.). Picking may be accomplished by providing a worker with a list of items and quantities, (i.e., a pick list). The worker then moves about the location gathering the items according to the pick list.
[0004] Various tools exist to improve the basic picking process. For example, a worker may use a mobile computing device (i.e., MCD) that is wirelessly connected to a warehouse management system (i.e., WMS) to receive and respond to picking instructions generated by the WMS. The worker's response may include information regarding the picking status (e.g., ready for next item), or may include validation information to help prevent picking errors (e.g., a check digit to confirm location).
[0005] Automated picking validation may be used in place of (or in addition to) user-enabled picking validation. For example, the worker may place items on a scale as part of a picking process. The weight of the picked item (or items) may be used to validate (i) that the correct item has been picked and/or (ii) that the correct quantity of the item has been picked. Despite these measures, picking errors still occur. Therefore, a need exists for a more accurate picking validation scheme that uses new and/or more sensors to verify one or more of an item's physical attributes automatically.
SUMMARY
[0006] Accordingly, in one aspect, the present invention embraces a method for validating a picking operation. The method begins with the picking of an item. The picked item is then placed on a platform in view of a sensor (or sensors). The sensor (or sensors) sense one or more sensed-physical attributes of the picked item. A computing device receives the sensed-physical attributes and compares these attributes to one or more expected-physical attributes stored on the computing device. If the one or more sensed-physical attributes do not match the one or more expected-physical attributes, then a picking-error alert is created.
[0007] In an exemplary embodiment of the method for validating a picking operation, the sensor (or sensors) includes a dimensioning system, and in this case, the one or more sensed-physical attributes can include a one-dimensional (1D), two-dimensional (2D), and/or three-dimensional (3D) measurement corresponding to the item's shape/size.
[0008] In another exemplary embodiment of the method for validating a picking operation, the sensors include a dimensioning system and a scale, and in this case, the one or more sensed-physical attributes can include the item's weight.
[0009] In another exemplary embodiment of the method for validating a picking operation, the sensors include a dimensioning system and a digital camera, and in this case, the one or more sensed-physical attributes can include the item's visual appearance.
[0010] In another exemplary embodiment of the method for validating a picking operation, the sensors include a dimensioning system, a scale, and a digital camera, and in this case, the sensed-physical attributes can include (i) a 1D, 2D, and/or 3D measurement of the item's shape/size, (ii) the item's weight, and (iii) the item's visual appearance.
[0011] In another exemplary embodiment of the method for validating a picking operation, the picking-error alert is an audio alert, a visual alert, a haptic alert, or some combination thereof.
[0012] In another exemplary embodiment of the method for validating a picking operation, the method includes updating the one or more expected-physical attributes using the one or more sensed-physical attributes if a match is found between the sensed-physical attributes and the expected-physical attributes. In some embodiments, the method may further include storing the updated expected-physical attributes on the computing device.
[0013] In another exemplary embodiment of the method for validating a picking operation, the method includes creating a picking-confirmation alert if the one or more sensed-physical attributes matches the one or more expected-physical attributes, and in this case, the picking-confirmation alert can be an audio alerL, a visual alerL, a hapLic alerL, or some combination thereof.
[0014] In another aspect, the present invention embraces a powered-industrial vehicle. The powered-industrial vehicle includes a user interface, at least one sensor, and a computing device, wherein the computing device is communicatively coupled to the user interface and the at least one sensor. The powered-industrial vehicle's at least one sensor is configured to (i) sense the physical attributes of an item placed on the power-industrial vehicle and (ii) output physical data. The powered-industrial-vehicle's computing-device receives the physical data from the at least one sensor and generates a set of sensed-physical attributes for the item. The computing device then retrieves a set of expected-physical attributes for the item from the computing device's memory. The set of sensed-physical attributes is compared to the set of expected-physical attributes, and based on the comparison, the computing device transmits an alert message to the user interface.
[0015] In an exemplary embodiment of the power-industrial vehicle, the at least one sensor includes a dimensioning system.
[0016] In another exemplary embodiment of the powered-industrial vehicle, the alert message includes (i) an indication that the wrong item has been placed on the powered-industrial vehicle, or (ii) an indication that the correct item has been placed on the powered-industrial vehicle.
[0017] In another aspect, the present invention embraces a picking system. The picking system includes a host computer that is communicatively coupled to a mobile computer. The host computer creates and sends a pick list to the mobile computer. The pick list includes one or more items to be picked and a set of expected physical attributes corresponding to the one or more items to be picked. The mobile computer is communicatively coupled to at least one sensor for sensing the physical attributes of a picked item and for outputting physical data. The mobile computer also includes a user interface, a memory for storing the pick list, and a processor communicatively coupled to the memory, the user interface, and the at least one sensor. The processor is configured by software to validate the picked item based on a set sensed-physical attributes generated using the physical data. Validation includes retrieving a set of expected-physical attributes for the picked item from memory and comparing the set of expected-physical attributes to the set of sensed-physical attributes. Based on this comparison, an alert message is transmitted to the user interface.
[0018] In an exemplary embodiment of the picking system, the at least one sensor includes a dimensioning system.
[0019] In another exemplary embodiment of the picking system, the at least one sensor includes (i) a dimensioning system and (ii) a scale and/or a digital camera.
[0020] In another exemplary embodiment of the picking system, the processor is further configured to update the set of expected-physical attributes using the set of sensed-physical attributes.
[0021] The foregoing illustrative summary, as well as other exemplary objectives and/or advantages of the invention, and the manner in which the same are accomplished, are further explained within the following detailed description and its accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
[0022] Figure 1 is a flow diagram illustrating a method for validating a picking operation according to an embodiment of the present invention.
[0023] Figure 2 graphically depicts the implementation of a powered-industrial vehicle according to a first exemplary embodiment of the present invention.
[0024] Figure 3 graphically depicts the implementation of a powered-industrial vehicle according a second exemplary embodiment of the present invention.
[0025] Figure 4 is a block diagram illustrating a picking system according to an exemplary embodiment of the present invention.
DETAILED DESCRIPTION
[0026] The present invention embraces an automatic picking validation system/method based on sensing physical aspects of a picked item and providing feedback (i.e., messages, alerts, etc.) to a user based on a comparison between the sensed-physical aspects and expected-physical aspects stored in memory.
[0027] Picking includes gathering particular quantities of items in an item-storage location (e.g., a warehouse). Typically, picking is performed as part of a logistics (e.g., supply-chain) process. Picking may be accomplished in a variety of ways (e.g., picker-to-part, part-to-picker, etc.), and different tools have been created to facilitate aspects of this work. These tools may help (1) picking management (e.g., warehouse management system (NMS)), (ii) directing a worker (e.g., mobile computers, voice headsets, etc.), (iii) lifting/carrying items (e.g., powered industrial vehicles), and (iv) sorting/counting items (e.g., robotics). Despite, this automation/mechanization picking is still a labor-intensive activity and prone to human error.
[0028] Picking errors are costly and should be avoided. To this end, workflows are often designed to eliminate picking errors by requiring workers to interact with the TAIMS (e.g., to confirm a location, to confirm number of items picked, etc.). Even with these measures, picking errors may still occur.
[0029] Automated picking validation (e.g., picking error detection, picking confirmation, etc.) may support or replace worker-supported (i.e., user-enabled) picking validation. Automation relies on the fact that picked items have distinguishing physical attributes (e.g., weight, size/shape, color, markings, etc.). By sensing a picked item's physical attributes and comparing these sensed-physical attributes to physical characteristics that are expected (i.e., expected-physical attributes), picking errors may be detected or correct picking may be confirmed.
[0030] Scales have been used to determine the weight of picked items as they are gathered. Validating items by weight, while useful, has not completely replaced user-enabled validation. One reason for this is that the measured weight is not always accurate. Scales designed for heavy items may not accurately weigh lighter items. In addition, when weighing many items, each item having a weight that slightly deviates from an expected value, errors can result when the weight deviations for each item accumulate.
[0031] Sensors may be combined wiLh (or may replace) Lhe scale in order to improve picking validation and/or add functionality to the picking process. The set of sensors may perform a variety of functions (e.g., sense the physical attributes of items, count items, weigh/measure items, read markings on items, etc.). A computing device, with memory and software algorithms running on a processor, may aggregate the data from the sensors to provide a more accurate representation of the picked item and further automate the picking process.
[0032] Machine vision may be used for picking validation. Machine vision is the automatic visual inspection and analysis of items placed in front of an imaging device (e.g., digital camera). Images from the imaging device may be processed to detect an item by its appearance (e.g., color, reflectivity, shape, markings, etc.). The image processing may include variety of algorithms (e.g., image-stitching, filter, thresholding, pixel counting, segmentation, edge detection, color analysis, pattern recognition, optical-characterrecognition, etc.). The processor for image processing (e.g., image processing unit) may be separate from the imaging device or combined with the imaging device (e.g., smart camera). The imaging device may capture images of visible or invisible light and in some machine vision systems utilize special illumination (e.g., ultraviolet (UV), infrared (IR), etc.) to make visible certain aspects of an item (e.g., security marks).
[0033] Systems to measure (without contact) an item in three dimensions (i.e., dimensioning systems, dimensioners) may be used for picking validation. Dimensioning systems typically emit radiation (e.g., light, ultrasonic, x-ray, etc.) towards an item and detect the reflected radiation from the item in order to determine the size/shape of the item. Dimensioning systems may use various methods to probe the item (e.g., timeof-flight, triangulation, etc.).
[0034] In one dimensioning system embodiment, an IR pattern (i.e., point cloud) is projected onto an item. The reflected IR pattern is detected on an imaging device (e.g., range camera) and the position of the pattern elements in the point cloud image are measured and compared to a known pattern.
Offsets in the imaged pattern correspond to changes in range (i.e., depth). Measured ranges may then be used to compute the dimensions of the item or to create a 3D model of the item. This dimensioning system is typically characterized by a projector and a range camera in a stereoscopic arrangement. The processing of the point cloud image may be performed by the dimensioning system or by a computing device (e.g., mobile computer, host computer, etc.) that is in communication with the dimensioner.
[0035] By combining the outputs from a plurality of sensors, the picked item may be validated more accurately. An exemplary method for validating a picking operation is shown in Figure (Fig.) 1. First, an item is picked 10. The item may be picked automatically and brought to the user (e.g., sorting system and/or conveyor belt) or the user may move to the item and pick it manually. The item is then placed on a platform 20. The platform may be a designated area in the warehouse or may be part of a powered industrial vehicle. The platform supports the item so that it may be measured by at least one sensor. In one possible embodiment, the platform may be part of a scale to weigh the item. In another possible embodiment, the platform may be in the field of view of a digital camera (e.g., machine vision system). In another possible embodiment, the platform may be in the field of view of a dimensioning system. In still another possible embodiment, the platform may be part of a scale and in the field of view of a digital camera and a dimensioner. Various combinations of the sensors are within the scope of the invention.
[0036] After placing the item on the platform, the item is measured using one or more sensors 30 to collect one or more sensed-physical attributes 40. The sensed physical attributes are then compared to a set of expected-physical attributes 45 stored in a computer readable memory. The comparison is performed by a computing device (e.g., mobile computer, host computer, etc.) that is communicatively coupled (e.g., wireless or wired) to the one or more sensors and the computer readable memory. The computing device may be body worn, handheld, desktop, laptop, or may be integrated with a powered-industrial vehicle. Algorithms running on the computer's processor then compare the sensed and expected physical attributes 50. If the attributes do not match, then a picking-error alert is created 60. Alternatively, if the comparison results in a match, then a picking confirmation may be created 70. The picking-error alert and the picking-confirmation alert may be an audio (e.g., voice), visual (e.g., text/graphics on a screen, signal light, printed message, etc.), or haptic (e.g., vibration) messages. In addition, the alerts may use some combination of these messages (e.g., a flashing light and a vibration).
[0037] The expected-physical attributes for an item may be stored in the computing device's memory. In one possible embodiment, the expected attributes for all possible picked items are installed in the memory during a setup procedure. In another possible embodiment, the expected attributes are downloaded as needed from a host computer (e.g., warehouse management system) and stored locally in the computing device's memory during the picking process.
[0038] Expected-physical attributes may be acquired through preliminary training (i.e., supervised learning) and used until it is necessary to retrain (e.g., until a new item is added, until an item's attributes change, etc.). In many cases, however, it is desirable for the picking validation to be adaptable, and in these cases, updating the expected-physical attributes using the sensed-physical attributes (i.e., machine learning) is required. For example, if the one or more sensed-physical attributes does not match the expected-physical attributes (e.g., an item's package has changed) then the expected-physical attributes may be updated 80. In one possible embodiment, the sensed-physical attributes replace or add to the expected-physical attributes stored in memory. In another possible embodiment, the sensed-physical attributes are uploaded to a host computing system along with other sensed-physical attributes (e.g., collected by another user, collected at a different time, etc.) to help create new expected-physical attributes (i.e., unsupervised learning).
[0039] The picking process may require a worker to use a powered-industrial vehicle (e.g., forklift, motorized trucks, stock-chasers, stackers, lifts, etc.) to gather items. Here, the sensors, the computing device, and a user interface may be combined or integrated with the vehicle to facilitate picking validation. One possible implementation of a powered-industrial vehicle according to an embodiment of the present invention vehicle is shown in Fig. 2. The powered-industrial vehicle 100 has a platform 110 onto which an item 120 may be placed. When placed on the platform 110, the item 120 is within the field of views 135,145 of a dimensioning system 130 and a digital camera 140. The platform 110 may be part of a scale (not shown) so that items placed on the platform may be measured for weight. The dimensioning system, the digital camera, and the scale are communicatively coupled (e.g., wired or wireless) to a computing device (not shown) that is integrated within the vehicle 100. The computing device is also communicatively coupled to a user interface 150. The user interface, shown in Fig. 2, is a display (e.g., touch display) for presenting and receiving information to/from a user 160. While shown as a touch display, interfacing with a user may be accomplished in a variety ways (e.g., voice headset, printer, etc.).
[0040] Another possible implementation of a powered-industrial vehicle according to an embodiment of the present invention is shown in Fig. 3. Here, the user interface is a voice headset 180 and the computing device is a body-worn mobile computing device (MCD) 170. A wireless link (e.g. BLUETOOTH connects the headset 180 and the MCD 170. The MCD 170 is also wirelessly coupled to the dimensioning system 130, the digital camera 140, and the scale (not shown). The MCD 170 is also wirelessly coupled to a host computer 200 via a wireless network (e.g., WLAN, WIFI0a, etc.) 190. In this configuration, the host computer 200 may perform the majority (or all) of the process steps necessary for picking validation. The host computer may also communicate with the user 160 via voice commands via the headset 180 (e.g., picking instructions, alert messages, etc.).
[0041] Multiple workers using multiple vehicles may operate simultaneously in a location (e.g., warehouse). In this case, the host computer (e.g., warehouse management system) may direct and validate picking for each user. The host computer may create a pick list for each user and transmit the user's pick list in its entirety or item-by-item to the user as needed. A picklist is a list of each item to be gathered along with the quantity to be picked. The present invention embraces a pick list that also includes the expected-physical attributes for one or more items in the pick list. The pick list (or portions of the pick list) may be downloaded to the computing device and stored locally in the computing device's memory.
[0042] Fig. 4 graphically depicts a block diagram of an exemplary picking system. The picking system includes a host computer 200 for creating a pick list 210 including at least one item to be picked and a set of expected-physical attributes associated for each item. The host computer transmits the pick list to a mobile computer 250. The mobile computer stores the pick list in its memory (e.g., RAM, HDD, SSD, etc.). The mobile computer includes, or is otherwise communicatively coupled to, at least one sensor (e.g., dimensioning system 130, scale 240, digital camera 140, etc.) to measure the attributes of a picked item 120. Physical data from the sensors 130,240, 140 are communicated to the mobile computer's processor 220 (e.g., CPU, microprocessor, multi-core processor, ASIC, FPGA, etc.). The processor is configured by software stored in the memory 230 to generate, from the physical data, a set of sensed-physical attributes for the picked item. The processor then retrieves the expected-physical attributes for the picked item from the memory 230 and compares the sensed-physical attributes to the expected-physical attributes. Based on the comparison, the processor creates an alert message (e.g., picking error, picking confirmation, etc.) and transmits this message to the user interface 225 (e.g., display, voice headset, printer, etc.). In some cases, the processor updates the expected-physical attributes measured for the particular picked item. The updated expected-physical attributes may then be stored in the memory 230 and/or communicated to the host computer 200.
[0043] In another embodiment of the picking system, the processing necessary for validation (i.e., comparing the sensed-physical attributes to expected-physical attributes, creating an alert message, etc.) may be performed on the host computer 200.
[0044] In some cases, automatic picking validation may require user interaction. The user 160, for example, may be required to input a picked-item identifier (e.g., serial number, part number, location, etc.) via the user interface 223 so that the correct expected-physical attributes may be retrieved from memory. In operation, a user may input this identifier via a keyboard/number pad, via voice, or by scanning a barcode on the picked item using a barcode reader communicatively coupled to the mobile computer 250.
[0045] The results of the picking validation may be used for other purposes besides picking validation. For example, a warehouse management system may monitor the results of the picking validation to compute worker effectiveness, identify workflow problems, and/or compute volumetric weights of the picked items for shipping (e.g., compute shipping costs).
[0046] To supplement the present disclosure, this application incorporates entirely by reference the following commonly assigned patents, patent application publications, and patent applications: U.S. Patent No. 6,832,725; U.S. Patent No. 7,128,266; U.S. Patent No. 7,159,783; U.S. Patent No. 7,413,127; U.S. Patent No. 7,726,575; U.S. Patent No. 8,294,969; U.S. Patent No. 8,317,105; U.S. Patent No. 8,322,622; U.S. Patent No. 8,366,005; U.S. Patent No. 8,371,507; U.S. Patent No. 8,376,233; U.S. Patent No. 8,381,979; U.S. Patent No. 8,390,909; U.S. Patent No. 8,408,464; U.S. Patent No. 8,408,468; U.S. Patent No. 8,408,469; U.S. Patent No. 8,424,768; U.S. Patent No. 8,448,863; U.S. Patent No. 8,457,013; U.S. Patent No. 8,459,557; U.S. Patent No. 8,469,272; U.S. Patent No. 8,474,712; U.S. Patent No. 8,479,992; U.S. Patent No. 8,490,877; U.S. Patent No. 8,517,271; U.S. Patent No. 8,523,076; U.S. Patent No. 8,528,818; U.S. Patent No. 8,544,737; U.S. Patent No. 8,548,242; U.S. Patent No. 8,548,420; U.S. Patent No. 8,550,335; U.S. Patent No. 8,550,354; U.S. Patent No. 8,550,357; U.S. Patent No. 8,556,174; U.S. Patent No. 8,556,176; U.S. Patent No. 8,556,177; U.S. Patent No. 8,559,767; U.S. Patent No. 8,599,957; U.S. Patent No. 8,561,895; U.S. Patent No. 8,561,903; U.S. Patent No. 8,561,905; U.S. Patent No. 8,565,107; U.S. Patent No. 8,571,307; U.S. Patent No. 8,579,200; U.S. Patent No. 8,583,924; U.S. Patent No. 8,584,945; U.S. Patent No. 8,587,595; U.S. Patent No. 8,587,697; U.S. Patent No. 8,588,869; U.S. Patent No. 8,590,789; U.S. Patent No. 8,596,539; U.S. Patent No. 8,596,542; U.S. Patent No. 8,596,543; U.S. Patent No. 8,599,271; U.S. Patent No. 8,599,957; U.S. Patent No. 8,600,158; U.S. Patent No. 8,600,167; U.S. Patent No. 8,602,309; U.S. Patent No. 8,608,053; U.S. Patent No. 8,608,071; U.S. Patent No. 8,611,309; U.S. Patent No. 8,615,487; U.S. Patent No. 8,616,454; U.S. Patent No. 8,621,123; U.S. Patent No. 8,622,303; U.S. Patent No. 8,628,013; U.S. Patent No. 8,628,015; U.S. Patent No. 8,628,016; U.S. Patent No. 8,629,926; U.S. Patent No. 8,630,491; U.S. Patent No. 8,635,309; U.S. Patent No. 8,636,200; U.S. Patent No. 8,636,212; U.S. Patent No. 8,636,215; U.S. Patent No. 8,636,224; U.S. Patent No. 8,638,806; U.S. Patent No. 8,640,958; U.S. Patent No. 8,640,960; U.S. Patent No. 8,643,717; U.S. Patent No. 8,646,692; U.S. Patent No. 8,646,694; U.S. Patent No. 8,657,200; U.S. Patent No. 8,659,397; U.S. Patent No. 8,668,149; U.S. Patent No. 8,678,285; U.S. Patent No. 8,678,286; U.S. Patent No. 8,682,077; U.S. Patent No. 8,687,282; U.S. Patent No. 8,692,927; U.S. Patent No. 8,695,880; U.S. Patent No. 8,698,949; U.S. Patent No. 8,717,494; U.S. Patent No. 8,717,494; U.S. Patent No. 8,720,783; U.S. Patent No. 8,723,804; U.S. Patent No. 8,723,904; U.S. Patent No. 8,727,223; U.S. Patent No. D702,237; U.S. Patent No. 8,740,082; U.S. Patent No. 8,740,085; U.S. Patent No. 8,746,563; U.S. Patent No. 8,750,445; U.S. Patent No. 8,752,766; U.S. Patent No. 8,756,059; U.S. Patent No. 8,757,495; U.S. Patent No. 8,760,563; U.S. Patent No. 8,763,909; U.S. Patent No. 8,777,108; U.S. Patent No. 8,777,109; U.S. Patent No. 8,779,898; U.S. Patent No. 8,781,520; U.S. Patent No. 8,783,573; U.S. Patent No. 8,789,757; U.S. Patent No. 8,789,758; U.S. Patent No. 6,789,759; U.S. Patent No. 8,794,520; U.S. Patent No. 8,794,522; U.S. Patent No. 8,794,525; U.S. Patent No. 8,794,526; U.S. Patent No. 8,798,367; U.S. Patent No. 8,807,431; U.S. Patent No. 8,807,432; U.S. Patent No. 8,820,630; U.S. Patent No. 8,822,848; U.S. Patent No. 8,824,692; U.S. Patent No. 8,824,696; U.S. Patent No. 8,842,849; U.S. Patent No. 8,844,822; U.S. Patent No. 8,844,823; U.S. Patent No. 8,849,019; U.S. Patent No. 8,851,383; U.S. Patent No. 8,854,633; U.S. Patent No. 8,866,963; U.S. Patent No. 8,868,421; U.S. Patent No. 8,868,519; U.S. Patent No. 8,868,802; U.S. Patent No. 8,868,803; U.S. Patent No. 8,870,074; U.S. Patent No. 8,879,639; U.S. Patent No. 8,880,426; U.S. Patent No. 8,881,983; U.S. Patent No. 8,881,987; U.S. Patent No. 8,903,172; U.S. Patent No. 8,908,995; U.S. Patent No. 8,910,870; U.S. Patent No. 8,910,875; U.S. Patent No. 8,914,290; U.S. Patent No. 8,914,788; U.S. Patent No. 8,915,439; U.S. Patent No. 8,915,444; U.S. Patent No. 8,916,789; U.S. Patent No. 8,918,250; U.S. Patent No. 8,918,564; U.S. Patent No. 8,925,818; U.S. Patent No. 8,939,374; U.S. Patent No. 8,942,480; U.S. Patent No. 8,944,313; U.S. Patent No. 8,944,327; U.S. Patent No. 8,944,332; U.S. Patent No. 8,950,678; U.S. Patent No. 8,967,466; U.S. Patent No. 8,971,346; U.S. Patent No. 8,976,030; U.S. Patent No. 8,976,368; U.S. Patent No. 6,978,981; U.S. Patent No. 8,978,983; U.S. Patent No. 8,978,984; U.S. Patent No. 8,985,456; U.S. Patent No. 8,985,457; U.S. Patent No. 8,985,459; U.S. Patent No. 6,985,461; U.S. Patent No. 8,968,578; U.S. Patent No. 8,988,590; U.S. Patent No. 8,991,704; U.S. Patent No. 8,996,194; U.S. Patent No. 8,996,384; U.S. Patent No. 9,002,641; U.S. Patent No. 9,007,368; U.S. Patent No. 9,010,641; U.S. Patent No. 9,015,513; U.S. Patent No. 9,016,576; U.S. Patent No. 9,022,288; U.S. Patent No. 9,030,964; U.S. Patent No. 9,033,240; U.S. Patent No. 9,033,242; U.S. Patent No. 9,036,054; U.S. Patent No. 9,037,344; U.S. Patent No. 9,038,911; U.S. Patent No. 9,038,915; U.S. Patent No. 9,047,098; U.S. Patent No. 9,047,359; U.S. Patent No. 9,047,420; U.S. Patent No. 9,047,525; U.S. Patent No. 9,047,531; U.S. Patent No. 9,053,055; U.S. Patent No. 9,053,378; U.S. Patent No. 9,053,380; U.S. Patent No. 9,058,526; U.S. Patent No. 9,064,165; U.S. Patent No. 9,064,167; U.S. Patent No. 9,064,168; U.S. Patent No. 9,064,254; U.S. Patent No. 9,066,032; U.S. Patent No. 9,070,032; U.S. Design Patent No. D716,285; U.S. Design Patent No. D723,560; U.S. Design Patent No. D730,357; U.S. Design Patent No. D730,901; U.S. Design Patent No. D730,902 U.S. Design Patent No. D733,112; U.S. Design Patent No. D734,339; International Publication No. 2013/163789; International Publication No. 2013/173985; International Publication No. 2014/019130; International Publication No. 2014/110495; U.S. Patent Application Publication No. 2008/0185432; U.S. Patent Application Publication No. 2009/0134221; U.S. Patent Application Publication No. 2010/0177080; U.S. Patent Application Publication No. 2010/0177076; U.S. Patent Application Publication No. 2010/0177707; U.S. Patent Application Publication No. 2010/0177749; U.S. Patent Application Publication No. 2010/0265880; U.S. Patent Application Publication No. 2011/0202554; U.S. Patent Application Publication No. 2012/0111946; U.S. Patent Application Publication No. 2012/0168511; U.S. Patent Application Publication No. 2012/0168512; U.S. Patent Application Publication No. 2012/0193423; U.S. Patent Application Publication No. 2012/0203647; U.S. Patent Application Publication No. 2012/0223141; U.S. Patent Application Publication No. 2012/0228382; U.S. Patent Application Publication No. 2012/0248188; U.S. Patent Application Publication No. 2013/0043312; U.S. Patent Application Publication No. 2013/0082104; U.S. Patent Application Publication No. 2013/0175341; U.S. Patent Application Publication No. 2013/0175343; U.S. Patent Application Publication No. 2013/0257744; U.S. Patent Application Publication No. 2013/0257759; U.S. Patent Application Publication No. 2013/0270346; U.S. Patent Application Publication No. 2013/0287258; U.S. Patent Application Publication No. 2013/0292475; U.S. Patent Application Publication No. 2013/0292477; U.S. Patent Application Publication No. 2013/0293539; U.S. Patent Application Publication No. 2013/0293540; U.S. Patent Application Publication No. 2013/0306728; U.S. Patent Application Publication No. 2013/0306731; U.S. Patent Application Publication No. 2013/0307964; U.S. Patent Application Publication No. 2013/0308625; U.S. Patent Application Publication No. 2013/0313324; U.S. Patent Application Publication No. 2013/0313325; U.S. Patent Application Publication No. 2013/0342717; U.S. Patent Application Publication No. 2014/0001267; U.S. Patent Application Publication No. 2014/0008439; U.S. Patent Application Publication No. 2014/0025584; U.S. Patent Application Publication No. 2014/0034734; U.S. Patent Application Publication No. 2014/0036848; U.S. Patent Application Publication No. 2014/0039693; U.S. Patent Application Publication No. 2014/0042814; U.S. Patent Application Publication No. 2014/0049120; U.S. Patent Application Publication No. 2014/0049635; U.S. Patent Application Publication No. 2014/0061306; U.S. Patent Application Publication No. 2014/0063289; U.S. Patent Application Publication No. 2014/0066136; U.S. Patent Application Publication No. 2014/0067692; U.S. Patent Application Publication No. 2014/0070005; U.S. Patent Application Publication No. 2014/0071840; U.S. Patent Application Publication No. 2014/0074746; U.S. Patent Application Publication No. 2014/0076974; U.S. Patent Application Publication No. 2014/0078341; U.S. Patent Application Publication No. 2014/0078345; U.S. Patent Application Publication No. 2014/0097249; U.S. Patent Application Publication No. 2014/0098792; U.S. Patent Application Publication No. 2014/0100813; U.S. Patent Application Publication No. 2014/0103115; U.S. Patent Application Publication No. 2014/0104413; U.S. Patent Application Publication No. 2014/0104414; U.S. Patent Application Publication No. 2014/0104416; U.S. Patent Application Publication No. 2014/0104451; U.S. Patent Application Publication No. 2014/0106394; U.S. Patent Application Publication No. 2014/0106725; U.S. Patent Application Publication No. 2014/0108010; U.S. Patent Application Publication No. 2014/0108402; U.S. Patent Application Publication No. 2014/0110485; U.S. Patent Application Publication No. 2014/0114530; U.S. Patent Application Publication No. 2014/0124577; U.S. Patent Application Publication No. 2014/0124579; U.S. Patent Application Publication No. 2014/0125842; U.S. Patent Application Publication No. 2014/0125853; U.S. Patent Application Publication No. 2014/0125999; U.S. Patent Application Publication No. 2014/0129378; U.S. Patent Application Publication No. 2014/0131438; U.S. Patent Application Publication No. 2014/0131441; U.S. Patent Application Publication No. 2014/0131443; U.S. Patent Application Publication No. 2014/0131444; U.S. Patent Application Publication No. 2014/0131445; U.S. Patent Application Publication No. 2014/0131448; U.S. Patent Application Publication No. 2014/0133379; U.S. Patent Application Publication No. 2014/0136208; U.S. Patent Application Publication No. 2014/0140585; U.S. Patent Application Publication No. 2014/0151453; U.S. Patent Application Publication No. 2014/0152882; U.S. Patent Application Publication No. 2014/0158770; U.S. Patent Application Publication No. 2014/0159869; U.S. Patent Application Publication No. 2014/0166755; U.S. Patent Application Publication No. 2014/0166759; U.S. Patent Application Publication No. 2014/0168787; U.S. Patent Application Publication No. 2014/0175165; U.S. Patent Application Publication No. 2014/0175172; U.S. Patent Application Publication No. 2014/0191644; U.S. Patent Application Publication No. 2014/0191913; U.S. Patent Application Publication No. 2014/0197238; U.S. Patent Application Publication No. 2014/0197239; U.S. Patent Application Publication No. 2014/0197304; U.S. Patent Application Publication No. 2014/0214631; U.S. Patent Application Publication No. 2014/0217166; U.S. Patent Application Publication No. 2014/0217180; U.S. Patent Application Publication No. 2014/0231300; U.S. Patent Application Publication No. 2014/0232930; U.S. Patent Application Publication No. 2014/0247315; U.S. Patent Application Publication No. 2014/0263493; U.S. Patent Application Publication No. 2014/0263645; U.S. Patent Application Publication No. 2014/0267609; U.S. Patent Application Publication No. 2014/0270196; U.S. Patent Application Publication No. 2014/0270229; U.S. Patent Application Publication No. 2014/0278387; U.S. Patent Application Publication No. 2014/0278391; U.S. Patent Application Publication No. 2014/0282210; U.S. Patent Application Publication No. 2014/0284384; U.S. Patent Application Publication No. 2014/0288933; U.S. Patent Application Publication No. 2014/0297058; U.S. Patent Application Publication No. 2014/0299665; U.S. Patent Application Publication No. 2014/0312121; U.S. Patent Application Publication No. 2014/0319220; U.S. Patent Application Publication No. 2014/0319221; U.S. Patent Application Publication No. 2014/0326787; U.S. Patent Application Publication No. 2014/0332590; U.S. Patent Application Publication No. 2014/0344943; U.S. Patent Application Publication No. 2014/0346233; U.S. Patent Application Publication No. 2014/0351317; U.S. Patent Application Publication No. 2014/0353373; U.S. Patent Application Publication No. 2014/0361073; U.S. Patent Application Publication No. 2014/0361082; U.S. Patent Application Publication No. 2014/0362184; U.S. Patent Application Publication No. 2014/0363015; U.S. Patent Application Publication No. 2014/0369511; U.S. Patent Application Publication No. 2014/0374483; U.S. Patent Application Publication No. 2014/0374485; U.S. Patent Application Publication No. 2015/0001301; U.S. Patent Application Publication No. 2015/0001304; U.S. Patent Application Publication No. 2015/0003673; U.S. Patent Application Publication No. 2015/0009338; U.S. Patent Application Publication No. 2015/0009610; U.S. Patent Application Publication No. 2015/0014416; U.S. Patent Application Publication No. 2015/0021397; U.S. Patent Application Publication No. 2015/0028102; U.S. Patent Application Publication No. 2015/0028103; U.S. Patent Application Publication No. 2015/0028104; U.S. Patent Application Publication No. 2015/0029002; U.S. Patent Application Publication No. 2015/0032709; U.S. Patent Application Publication No. 2015/0039309; U.S. Patent Application Publication No. 2015/0039878; U.S. Patent Application Publication No. 2015/0040378; U.S. Patent Application Publication No. 2015/0048168; U.S. Patent Application Publication No. 2015/0049347; U.S. Patent Application Publication No. 2015/0051992; U.S. Patent Application Publication No. 2015/0053766; U.S. Patent Application Publication No. 2015/0053768; U.S. Patent Application Publication No. 2015/0053769; U.S. Patent Application Publication No. 2015/0060544; U.S. Patent Application Publication No. 2015/0062366; U.S. Patent Application Publication No. 2015/0063215; U.S. Patent Application Publication No. 2015/0063676; U.S. Patent Application Publication No. 2015/0069130; U.S. Patent Application Publication No. 2015/0071819; U.S. Patent Application Publication No. 2015/0083800; U.S. Patent Application Publication No. 2015/0086114; U.S. Patent Application Publication No. 2015/0088522; U.S. Patent Application Publication No. 2015/0096872; U.S. Patent Application Publication No. 2015/0099557; U.S. Patent Application Publication No. 2015/0100196; U.S. Patent Application Publication No. 2015/0102109; U.S. Patent Application Publication No. 2015/0115035; U.S. Patent Application Publication No. 2015/0127791; U.S. Patent Application Publication No. 2015/0128116; U.S. Patent Application Publication No. 2015/0129659; U.S. Patent Application Publication No. 2015/0133047; U.S. Patent Application Publication No. 2015/0134470; U.S. Patent Application Publication No. 2015/0136851; U.S. Patent Application Publication No. 2015/0136854; U.S. Patent Application Publication No. 2015/0142492; U.S. Patent Application Publication No. 2015/0144692; U.S. Patent Application Publication No. 2015/0144698; U.S. Patent Application Publication No. 2015/0144701; U.S. Patent Application Publication No. 2015/0149946; U.S. Patent Application Publication No. 2015/0161429; U.S. Patent Application Publication No. 2015/0169925; U.S. Patent Application Publication No. 2015/0169929; U.S. Patent Application Publication No. 2015/0178523; U.S. Patent Application Publication No. 2015/0176534; U.S. Patent Application Publication No. 2015/0178535; U.S. Patent Application Publication No. 2015/0178536; U.S. Patent Application Publication No. 2015/0178537; U.S. Patent Application Publication No. 2015/0181093; U.S. Patent Application Publication No. 2015/0181109; U.S. Patent Application No. 13/367,978 for a Laser Scanning Module Employing an Elastomeric U-Hinge Based Laser Scanning Assembly, filed February 7, 2012 (Feng et al.); U.S. Patent Application No. 29/458,405 for an Electronic Device, filed June 19, 2013 (Fitch et al.); U.S. Patent Application No. 29/459,620 for an Electronic Device Enclosure, filed July 2, 2013 (London et al.); U.S. Patent Application No. 29/468,118 for an Electronic Device Case, filed September 26, 2013 (Oberpriller et al.); U.S. Patent Application No. 14/150,393 for Indicia-reader Having Unitary Construction Scanner, filed January 8, 2014 (Colavito et al.); U.S. Patent Application No. 14/200,405 for Indicia Reader for Size-Limited Applications filed March 7, 2014 (Feng et al.); U.S. Patent Application No. 14/231,898 for Hand-Mounted Indicia-Reading Device with Finger Motion Triggering filed April 1, 2014 (Van Horn et al.); U.S. Patent Application No. 29/486,759 for an Imaging Terminal, filed April 2, 2014 (Oberpriller et al.); U.S. Patent Application No. 14/257,364 for Docking System and (Method Using Near Field Communication filed April 21, 2014 (Showering); U.S. Patent Application No. 14/264,173 for Autofocus Lens System for Indicia Readers filed April 29, 2014 (Ackley et a/.); U.S. Patent Application No. 14/277,337 for MULTIPURPOSE OPTICAL READER, filed May 14, 2014 (Jovanovski et al.); U.S. Patent Application No. 14/283,282 for TERMINAL HAVING ILLUMINATION AND FOCUS CONTROL filed May 21, 2014 (Liu et al.); U.S. Patent Application No. 14/327,827 for a MOBILE-PHONE ADAPTER FOR ELECTRONIC TRANSACTIONS, filed July 10, 2014 (Hejl); U.S. Patent Application No. 14/334,934 for a SYSTEM AND METHOD FOR INDICIA VERIFICATION, filed July 18, 2014 (iejl); U.S. Patent Application No. 14/339,708 for LASER SCANNING CODE SYMBOL READING SYSTEM, filed July 24, 2014 (Xian et al.); U.S. Patent Application No. 14/340,627 for an AXIALLY REINFORCED FLEXIBLE SCAN ELEMENT, filed July 25, 2014 (Rueblinger et al.); U.S. Patent Application No. 14/446,391 for MULTIFUNCTION POINT OF SALE APPARATUS WITH OPTICAL SIGNATURE CAPTURE filed July 30, 2014 (Good et al.); U.S. Patent Application No. 14/452,697 for INTERACTIVE INDICIA READER, filed August 6, 2014 (Todeschini); U.S. Patent Application No. 14/453,019 for DIMENSIONING SYSTEM WITH GUIDED ALIGNMENT, filed August 6, 2014 (Li et al.); U.S. Patent Application No. 14/462,801 for MOBILE COMPUTING DEVICE WITH DATA COGNITION SOFTWARE, filed on August 19, 2014 (Todeschini et al.); U.S. Patent Application No. 14/483,056 for VARIABLE DEPTH OF FIELD BARCODE SCANNER filed Sep. 10, 2014 (McCloskey et al.); U.S. Patent Application No. 14/513,808 for IDENTIFYING INVENTORY ITEMS IN A STORAGE FACILITY filed Oct. 14, 2014 (Singel et al.); U.S. Patent Application No. 14/519,195 for HANDHELD DIMENSIONING SYSTEM WITH FEEDBACK filed Oct. 21, 2014 (Laffargue et al.); U.S. Patent Application No. 14/519,179 for DIMENSIONING SYSTEM WITH MULTIPATH INTERFERENCE MITIGATION filed Oct. 21, 2014 (Thuries et al.); U.S. Patent Application No. 14/519,211 for SYSTEM AND METHOD FOR DIMENSIONING filed Oct. 21, 2014 (Ackley et al.); U.S. Patent Application No. 14/519,233 for HANDHELD DTMENSIONFR WITH DATA-QUALITY INDICATION filed Oct. 21, 2014 (Laffargue et al.); U.S. Patent Application No. 14/519,249 for HANDHELD DIMENSIONING SYSTEM WITH MEASUREMENT-CONFORMANCE FEEDBACK filed Oct. 21, 2014 (Ackley et al.); U.S. Patent Application No. 14/527,191 for METHOD AND SYSTEM FOR RECOGNIZING SPEECH USING WILDCARDS IN AN EXPECTED RESPONSE filed Oct. 29, 2014 (Braho et al.); U.S. Patent Application No. 14/529,563 for ADAPTABLE INTERFACE FOR A MOBILE COMPUTING DEVICE filed Oct. 31, 2014 (Schoon et al.); U.S. Patent Application No. 14/529,857 for BARCODE READER WITH SECURITY FEATURES filed October 31, 2014 (Todeschini et al.); U.S. Patent Application No. 14/398,542 for PORTABLE ELECTRONIC DEVICES HAVING A SEPARATE LOCATION TRIGGER UNIT FOR USE IN CONTROLLING AN APPLICATION UNIT filed November 3, 2014 (Sian et al.); U.S. Patent Application No. 14/531,154 for DIRECTING AN INSPECTOR THROUGH AN INSPECTION filed Nov. 3, 2014 (Miller et al.); U.S. Patent Application No. 14/533,319 for BARCODE SCANNING SYSTEM USING WEARABLE DEVICE WITH EMBEDDED CAMERA filed Nov. 5, 2014 (Todeschini); U.S. Patent Application No. 14/535,764 for CONCATENATED EXPECTED RESPONSES FOR SPEECH RECOGNITION filed Nov. 7, 2014 (Braho et al.); U.S. Patent Application No. 14/568,305 for AUTO-CONTRAST VIEWFINDER FOR AN INDICIA READER filed Dec. 12, 2014 (Todeschini); U.S. Patent Application No. 14/573,022 for DYNAMIC DIAGNOSTIC INDICATOR GENERATION filed Dec. 17, 2014 (Goldsmith); U.S. Patent Application No. 14/578,627 for SAFETY SYSTEM AND METHOD filed Dec. 22, 2014 (Ackley et al.); U.S. Patent Application No. 14/580,262 for MEDIA GATE FOR THERMAL TRANSFER PRINTERS filed Dec. 23, 2014 (Bowles); U.S. Patent Application No. 14/590,024 for SHELVING AND PACKAGE LOCATING SYSTEMS FOR DELIVERY VEHICLES filed January 6, 2015 (Payne); U.S. Patent Application No. 14/596,757 for SYSTEM AND METHOD FOR DETECTING BARCODE PRINTING ERRORS filed Jan. 14, 2015 (Ackley); U.S. Patent Application No. 14/416,147 for OPTICAL READING APPARATUS HAVING VARIABLE SETTINGS filed January 21, 2015 (Chen et al.); U.S. Patent Application No. 14/614,706 for DEVICE FOR SUPPORTING AN ELECTRONIC TOOL ON A USER'S HAND filed Feb. 5, 2015 (Oberpriller et al.); U.S. Patent Application No. 14/614,796 for CARGO APPORTIONMENT TECHNIQUES filed Feb. 5, 2015 (Morton et al.); U.S. Patent Application No. 29/516,892 for TABLE COMPUTER filed Feb. 6, 2015 (Bidwell et al.); U.S. Patent Application No. 14/619,093 for METHODS FOR TRAINING A SPEECH RECOGNITION SYSTEM filed Feb. 11, 2015 (Pecorari); U.S. Patent Application No. 14/628,708 for DEVICE, SYSTEM, AND METHOD FOR DETERMINING THE STATUS OF CHECKOUT LANES filed Feb. 23, 2015 (Todeschini); U.S. Patent Application No. 14/630,841 for TERMINAL INCLUDING IMAGING ASSEMBLY filed Feb. 25, 2015 (Gomez et al.); U.S. Patent Application No. 14/635,346 for SYSTEM AND METHOD FOR RELIABLE STORE-AND-FORWARD DATA HANDLING BY ENCODED INFORMATION READING TERMINALS filed March 2, 2015 (Sevier); U.S. Patent Application No. 29/519,017 for SCANNER filed March 2, 2015 (Zhou et al.); U.S. Patent Application No. 14/405,278 for DESIGN PATTERN FOR SECURE STORE filed March 9, 2015 (Zhu et al.); U.S. Patent Application No. 14/660,970 for DECODABLE INDICIA READING TERMINAL WITH COMBINED ILLUMINATION filed March 18, 2015 (Kearney et al.); U.S. Patent Application No. 14/661,013 for REPROGRAMMING SYSTEM AND METHOD FOR DEVICES INCLUDING PROGRAMMING SYMBOL filed March 18, 2015 (Soule et al.); U.S. Patent Application No. 14/662,922 for MULTIFUNCTION POINT OF SALE SYSTEM filed March 19, 2015 (Van Horn et al.); U.S. Patent Application No. 14/663,638 for VEHICLE MOUNT COMPUTER WITH CONFIGURABLE IGNITION SWITCH BEHAVIOR filed March 20, 2015 (Davis et al.); U.S. Patent Application No. 14/664,063 for METHOD AND APPLICATION FOR SCANNING A BARCODE WITH A SMART DEVICE WHILE CONTINUOUSLY RUNNING AND DISPLAYING AN APPLICATION ON THE SMART DEVICE DISPLAY filed March 20, 2015 (Todeschini); U.S. Patent Application No. 14/669,280 for TRANSFORMING COMPONENTS OF A WEB PAGE TO VOICE PROMPTS filed March 26, 2015 (Funyak et al.); U.S. Patent Application No. 14/674,329 for AMER FOR BARCODE SCANNING filed March 31, 2015 (Bidwell); U.S. Patent Application No. 14/676,109 for INDICIA READER filed April 1, 2015 (Huck); U.S. Patent Application No. 14/676,327 for DEVICE MANAGEMENT PROXY FOR SECURE DEVICES filed April 1, 2015 (Yeakley et al.); U.S. Patent Application No. 14/676,898 for NAVIGATION SYSTEM CONFIGURED TO INTEGRATE MOTION SENSING DEVICE INPUTS filed April 2, 2015 (Showering); U.S. Patent Application No. 14/679,275 for DIMENSIONING SYSTEM CALIBRATION SYSTEMS AND METHODS filed April 6, 2015 (Laffargue et al.); U.S. Patent Application No. 29/523,098 for HANDLE FOR A TABLET COMPUTER filed April 7, 2015 (Bidwell et al.); U.S. Patent Application No. 14/682,615 for SYSTEM AND METHOD FOR POWER MANAGEMENT OF MOBILE DEVICES filed April 9, 2015 (Murawski et al.); U.S. Patent Application No. 14/686,822 for MULTIPLE PLATFORM SUPPORT SYSTEM AND METHOD filed April 15, 2015 (Qu et al.); U.S. Patent Application No. 14/687,289 for SYSTEM FOR COMMUNICATION VIA A PERIPHERAL HUB filed April 15, 2015 (Kohtz et al.); U.S. Patent Application No. 29/524,186 for SCANNER filed April 17, 2015 (Zhou et al.); U.S. Patent Application No. 14/695,364 for MEDICATION MANAGEMENT SYSTEM filed April 24, 2015 (Sewell et al.); U.S. Patent Application No. 14/695,923 for SECURE UNATTENDED NETWORK AUTHENTICATION filed April 24, 2015 (Kubler et al.); U.S. Patent Application No. 29/525,068 for TABLET COMPUTER WITH REMOVABLE SCANNING DEVICE filed April 27, 2013 (Schulte et al.); U.S. Patent Application No. 14/699,436 for SYMBOL READING SYSTEM HAVING PREDICTIVE DIAGNOSTICS filed April 29, 2015 (Nahill et al.); U.S. Patent Application No. 14/702,110 for SYSTEM AND METHOD FOR REGULATING BARCODE DATA INJECTION INTO A RUNNING APPLICATION ON A SMART DEVICE filed May 1, 2015 (Todeschini et al.); U.S. Patent Application No. 14/702,979 for TRACKING BATTERY CONDITIONS filed May 4, 2015 (Young et al.); U.S. Patent Application No. 14/704,050 for INTERMEDIATE LINEAR POSITIONING filed May 5, 2015 (Charpentier et al.); U.S. Patent Application No. 14/705,012 for HANDS-FREE HUMAN MACHINE INTERFACE RESPONSIVE TO A DRIVER OF A VEHICLE filed May 6, 2015 (Fitch et al.); U.S. Patent Application No. 14/705,407 for METHOD AND SYSTEM TO PROTECT SOFTWARE-BASED NETWORK-CONNECTED DEVICES FROM ADVANCED PERSISTENT THREAT filed May 6, 2015 (Hussey et al.); U.S. Patent Application No. 14/707,037 for SYSTEM AND METHOD FOR DISPLAY OF INFORMATION USING A VEHICLE-MOUNT COMPUTER filed May 8, 2015 (Chamberlin); U.S. Patent Application No. 14/707,123 for APPLICATION INDEPENDENT DFX/UCS INTERFACE filed May 8, 2015 (Pape); U.S. Patent Application No. 14/707,492 for METHOD AND APPARATUS FOR READING OPTICAL INDICIA USING A PLURALITY OF DATA SOURCES filed May 8, 2015 (Smith et al.); U.S. Patent Application No. 14/710,666 for PRE-PAID USAGE SYSTEM FOR ENCODED INFORMATION READING TERMINALS filed May 13, 2015 (Smith); U.S. Patent Application No. 29/526,918 for CHARGING BASE filed May 14, 2015 (Fitch et al.); U.S. Patent Application No. 14/715,672 for AUGUMENTED REALITY ENABLED HAZARD DISPLAY filed May 19, 2015 (Venkatesha et al.); U.S. Patent Application No. 14/715,916 for EVALUATING IMAGE VALUES filed May 19, 2015 (Ackley); U.S. Patent Application No. 14/722,608 for INTERACTIVE USER INTERFACE FOR CAPTURING A DOCUMENT IN AN IMAGE SIGNAL filed May 27, 2015 (Showering et al.); U.S. Patent Application No. 29/528,165 for IN-COUNTER BARCODE SCANNER filed May 27, 2015 (Oberpriller et al.); U.S. Patent Application No. 14/724,134 for ELECTRONIC DEVICE WITH WIRELESS PATH SELECTION CAPABILITY filed May 28, 2015 (Wang et al.); U.S. Patent Application No. 14/724,849 for METHOD OF PROGRAMMING THE DEFAULT CABLE INTERFACE SOFTWARE IN AN INDICIA READING DEVICE filed May 29, 2015 (Barten); U.S. Patent Application No. 14/724,908 for IMAGING APPARATUS HAVING IMAGING ASSEMBLY filed May 29, 2015 (Barber et al.); U.S. Patent Application No. 14/725,352 for APPARATUS AND METHODS FOR MONITORING ONE OR MORE PORTABLE DATA TERMINALS (Caballero et al.); U.S. Patent Application No. 29/528,590 for ELECTRONIC DEVICE filed May 29, 2015 (Fitch et al.); U.S. Patent Application No. 29/528,890 for MOBILE COMPUTER HOUSING filed June 2, 2015 (Fitch et al.); U.S. Patent Application No. 14/728,397 for DEVICE MANAGEMENT USING VIRTUAL INTERFACES CROSS-REFERENCE TO RELATED APPLICATIONS filed June 2, 2015 (Caballero); U.S. Patent Application No. 14/732,870 for DATA COLLECTION MODULE AND SYSTEM filed June 8, 2015 (Powilleit); U.S. Patent Application No. 29/529,441 for INDICIA READING DEVICE filed June 8, 2015 (Zhou et al.); U.S. Patent Application No. 14/735,717 for INDICIA-READING SYSTEMS HAVING AN INTERFACE WITH A USER'S NERVOUS SYSTEM filed June 10, 2015 (Todeschini); U.S. Patent Application No. 14/738,038 for METHOD OF AND SYSTEM FOR DETECTING OBJECT WEIGHING INTERFERENCES filed June 12, 2015 (Amundsen et al.); U.S. Patent Application No. 14/740,320 for TACTILE SWITCH FOR A MOBILE ELECTRONIC DEVICE filed June 16, 2015 (Bandringa); U.S. Patent Application No. 14/740,373 for CALIBRATING A VOLUME DIMENSIONER filed June 16, 2015 (Ackley et al.); U.S. Patent Application No. 14/742,818 for INDICIA READING SYSTEM EMPLOYING DIGITAL GAIN CONTROL filed June 1E, 2015 (Xian et al.); U.S. Patent Application No. 14/743,257 for WIRELESS MESH POINT PORTABLE DATA TERMINAL filed June 18, 2015 (Wang et al.); U.S. Patent Application No. 29/530,600 for CYCLONE filed June 16, 2015 (Vargo et al); U.S. Patent Application No. 14/744,633 for IMAGING APPARATUS COMPRISING IMAGE SENSOR ARRAY HAVING SHARED GLOBAL SHUTTER CIRCUITRY filed June 19, 2015 (Wang); U.S. Patent Application No. 14/744,836 for CLOUD-BASED SYSTEM FOR READING OF DECODABLE INDICIA filed June 19, 2015 (Todeschini et al.); U.S. Patent Application No. 14/745,006 for SELECTIVE OUTPUT OF DECODED MESSAGE DATA filed June 19, 2015 (Todeschini et al.); U.S. Patent Application No. 14/747,197 for OPTICAL PATTERN PROJECTOR filed June 23, 2015 (Thuries et al.); U.S. Patent Application No. 14/747,490 for DUAL-PROJECTOR THREE-DIMENSIONAL SCANNER filed June 23, 2015 (Jovanovski et al.);and U.S. Patent Application No. 14/748,446 for CORDLESS INDICIA READER WITH A MULTIFUNCTION COIL FOR WIRELESS CHARGING AND EAS DEACTIVATION, filed June 24, 2015 (Xie et al.).
[0047] In the specification and/or figures, typical embodiments of the invention have been disclosed. The present invention is not limited to such exemplary embodiments. The use of the term "and/or" includes any and all combinations of one or more of the associated listed items. The figures are schematic representations and so are not necessarily drawn to scale. Unless otherwise noted, specific terms have been used in a generic and descriptive sense and not for purposes of limitation.

Claims (23)

  1. CLAIMS1. A method for validating a picking operation, comprising: picking an item; placing the picked item on a platform in view of at least one sensor communicatively coupled to a computing device; sensing, using the at least one sensor, one or more sensed-physical attributes of the picked item; comparing, using the computing device, the one or more sensed-physical attributes to one or more expected-physical attributes stored on the computing device; and if the one or more sensed-physical attributes do not match the one or more expected-physical attributes, then creating a picking-error alert.
  2. 2. The method according to claim 1, wherein the at least one sensor comprises a dimensioning system.
  3. 3. The method according to claim 2, wherein the one or more sensed-physical attributes comprises a one-dimensional, two-dimensional, and/or three-dimensional measurement corresponding to the item's shape/size.
  4. 4. The method according to claim 2, wherein the at least one sensor comprises a scale.
  5. 5. The method according to claim 4, wherein the one or more sensed-physical attributes comprises the item's weight.
  6. 6. The method according to claim 2, wherein the at least one sensor comprises a digital camera.
  7. 7. The method according to claim 6, wherein the one or more sensed-physical attributes comprises the item's visual appearance.
  8. 8. The method according to claim 1, wherein the at least one sensor comprises a dimensioning system, a scale, and a digital camera.
  9. 9. The method according to claim 6, wherein the one or more sensed-physical attributes comprises (i) a one-dimensional, two-dimensional, and/or three-dimensional measurement corresponding to the item's shape/size, (ii) the item's weight, and (iii) the item's visual appearance.
  10. 10. The method according to claim 1, wherein the picking-error alert is (i) an audio alert, a visual alert, or a haptic alert or (ii) combinations thereof.
  11. 11. The method according to claim 1, further comprising: If the one or more sensed-physical attributes matches the one or more expected-physical attributes, then updating the one or more expected-physical attributes using the one or more sensed-physical attributes, and storing the updated one or more expected-physical attributes on the computing device.
  12. 12. The method according to claim 1, further comprising: if the one or more sensed-physical attributes matches the one or more expected-physical attributes, then creating a picking-confirmation alert.
  13. 13. The method according to claim 12, wherein the picking-confirmation alert is (i) an audio alert, a visual alert, or a haptic alert or (ii) combinations thereof.
  14. 14. A powered-industrial vehicle, comprising: a user interface; at least one sensor configured to: (i) sense physical attributes of an item placed on the powered-industrial vehicle and (ii) output physical data; and a computing device communicatively coupled to the user interface and the at least one sensor, wherein the computing device is configured to: (i) generate a set of sensed-physical attributes for the item based on the physical data, (ii) retrieve a set of expected-physical attributes for the item from the computing device's memory, (ii) compare the set of sensed-physical attributes to the set of expected-physical attributes, and (iii) based on the comparison, transmit an alert message to the user interface.
  15. 15. The powered-industrial vehicle according to claim 14, wherein the at least one sensor comprises a dimensioning system.
  16. 16. The powered-industrial vehicle according to claim 14, wherein the alert message comprises (i) an indication that the wrong item has been placed on the powered-industrial vehicle or (ii) an indication that the correct item has been placed on the powered-industrial vehicle.
  17. 17. A picking system, comprising: a host computer for creating a pick list, the pick list comprising one or more items to be picked and a set of expected-physical attributes corresponding to the one or more items to be picked; at least one sensor configured to: (i) sense physical attributes of a picked item and (ii) output physical data; and a mobile computer communicatively coupled to the host computer and the at least one sensor, wherein the mobile computer comprises: a user interface, a memory for storing the pick list, and a processor communicatively coupled to the memory, the user interface, and the at least one sensor, wherein the processor is configured by software to: (i) generate, based on the physical data, a set of sensed-physical attributes corresponding to the picked item, (ii) retrieve the set of expected-physical attributes for the picked item from the memory, (ii) compare the set of sensed-physical attributes to the set of expected-physical attributes, (iii) based on the comparison, create an alert message, and (iv) transmit the alert message to the user interface.
  18. 18. The picking system according to claim 17, wherein the at least one sensor comprises a dimensioning system.
  19. 19. The picking system according to claim 18, wherein the at least one sensor further comprises a scale and/or a digital camera.
  20. 20. The picking system according to claim 17, wherein the processor is further configured to update the set of expected-physical attributes using the set of sensed-physical attributes.
  21. 21. A method substantially as hereinbefore described with reference to and/or as illustrated in any one or more of the Figures.
  22. 22. A powered-industrial vehicle as hereinbefore described with reference to and/or as illustrated in any one or more of the Figures.
  23. 23. A picking system as hereinbefore described with reference to and/or as illustrated in any one or more of the Figures.
GB1517842.9A 2014-10-10 2015-10-08 System and method for picking validation Active GB2533676B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201462062175P 2014-10-10 2014-10-10
US14/865,797 US10810715B2 (en) 2014-10-10 2015-09-25 System and method for picking validation

Publications (3)

Publication Number Publication Date
GB201517842D0 GB201517842D0 (en) 2015-11-25
GB2533676A true GB2533676A (en) 2016-06-29
GB2533676B GB2533676B (en) 2019-07-03

Family

ID=55130787

Family Applications (1)

Application Number Title Priority Date Filing Date
GB1517842.9A Active GB2533676B (en) 2014-10-10 2015-10-08 System and method for picking validation

Country Status (1)

Country Link
GB (1) GB2533676B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111837147A (en) * 2018-03-12 2020-10-27 罗伯特·博世有限公司 Method and camera system for monitoring a sorting or packaging process

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE3335760A1 (en) * 1983-10-01 1985-04-25 Sachße, Lothar, 8500 Nürnberg Device in a mobile commissioning system
JPH04129902A (en) * 1990-09-20 1992-04-30 Nec Software Ltd Merchandise picking system
US5730252A (en) * 1995-02-27 1998-03-24 L'oreal Order preparation method, an order picking cart and order preparation system
US20110301994A1 (en) * 2010-06-07 2011-12-08 Tieman James K Wireless put-to-light system and method
JP2014210646A (en) * 2013-04-18 2014-11-13 三菱化学エンジニアリング株式会社 Commodity check system, commodity check device, and commodity check method
US20140350710A1 (en) * 2013-05-24 2014-11-27 Amazon Technologies, Inc. Tote Based Item Tracking

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
ATE540884T1 (en) * 2006-05-02 2012-01-15 Habitat Italiana S R L DEVICE FOR STORING AND RECEIVING OBJECTS OF DIFFERENT DIMENSIONS AND WEIGHTS, IN PARTICULAR BOOKS
US8201737B1 (en) * 2009-07-21 2012-06-19 Amazon Technologies, Inc. Validating item placement

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE3335760A1 (en) * 1983-10-01 1985-04-25 Sachße, Lothar, 8500 Nürnberg Device in a mobile commissioning system
JPH04129902A (en) * 1990-09-20 1992-04-30 Nec Software Ltd Merchandise picking system
US5730252A (en) * 1995-02-27 1998-03-24 L'oreal Order preparation method, an order picking cart and order preparation system
US20110301994A1 (en) * 2010-06-07 2011-12-08 Tieman James K Wireless put-to-light system and method
JP2014210646A (en) * 2013-04-18 2014-11-13 三菱化学エンジニアリング株式会社 Commodity check system, commodity check device, and commodity check method
US20140350710A1 (en) * 2013-05-24 2014-11-27 Amazon Technologies, Inc. Tote Based Item Tracking

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111837147A (en) * 2018-03-12 2020-10-27 罗伯特·博世有限公司 Method and camera system for monitoring a sorting or packaging process

Also Published As

Publication number Publication date
GB2533676B (en) 2019-07-03
GB201517842D0 (en) 2015-11-25

Similar Documents

Publication Publication Date Title
US10810715B2 (en) System and method for picking validation
US10897940B2 (en) Gloves having measuring, scanning, and displaying capabilities
US10049290B2 (en) Industrial vehicle positioning system and method
US10464349B2 (en) Method and system to calculate line feed error in labels on a printer
CN206601173U (en) Use view data calculated weight
US10775165B2 (en) Methods for improving the accuracy of dimensioning-system measurements
US10134112B2 (en) System and process for displaying information from a mobile computer in a vehicle
US10176521B2 (en) Augmented reality virtual product for display
US10552786B2 (en) Product and location management via voice recognition
US9729744B2 (en) System and method of border detection on a document and for producing an image of the document
US9864891B2 (en) Automatic print speed control for indicia printer
US9898635B2 (en) Point-of-sale (POS) code sensing apparatus
EP3163497A1 (en) Image transformation for indicia reading
EP3547237A1 (en) Confirming product location using a subset of a product identifier
US10909708B2 (en) Calibrating a dimensioner using ratios of measurable parameters of optic ally-perceptible geometric elements
US10183506B2 (en) Thermal printer having real-time force feedback on printhead pressure and method of using same
US10810541B2 (en) Methods for pick and put location verification
GB2533676A (en) System and method for picking validation
EP3246863A1 (en) Systems and methods for reducing picking operation errors
US10152664B2 (en) Backlit display detection and radio signature recognition
US9805257B1 (en) Printer method and apparatus
US20180196979A1 (en) Waking system in barcode scanner
US11157869B2 (en) Monitoring worker movement in a warehouse setting
EP3136048A1 (en) Gloves having measuring, scanning, and displaying capabilities
GB2535620A (en) Industrial vehicle positioning system and method