US20160224865A9 - Methods and systems for enabling vision based inventory management - Google Patents

Methods and systems for enabling vision based inventory management Download PDF

Info

Publication number
US20160224865A9
US20160224865A9 US13/945,403 US201313945403A US2016224865A9 US 20160224865 A9 US20160224865 A9 US 20160224865A9 US 201313945403 A US201313945403 A US 201313945403A US 2016224865 A9 US2016224865 A9 US 2016224865A9
Authority
US
United States
Prior art keywords
component
selecting
components
image
features
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US13/945,403
Other versions
US9665794B2 (en
US20140023276A1 (en
Inventor
Harikrishna Gandhinagara Narayana Rai
Krishnamurthy Sai Deepak
Radha Krishna Pisipati
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Infosys Ltd
Original Assignee
Infosys Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Infosys Ltd filed Critical Infosys Ltd
Publication of US20140023276A1 publication Critical patent/US20140023276A1/en
Assigned to Infosys Limited reassignment Infosys Limited ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RAI, HARIKRISHNA GANDHINAGARA NARAYANA, DEEPAK, KRISHNAMURTHY SAI, PISIPATI, RADHA KRISHNA
Publication of US20160224865A9 publication Critical patent/US20160224865A9/en
Application granted granted Critical
Publication of US9665794B2 publication Critical patent/US9665794B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06K9/6202
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/142Image acquisition using hand-held instruments; Constructional details of the instruments

Definitions

  • the invention relates generally to a method and system for inventory management.
  • it relates to a method and system for locating the position of components in store keeping unit for computer assisted inventory placement and auditing in manufacturing plant.
  • RFID technology is another alternate technology widely used for component and asset tracking in industries.
  • RFID based solutions have several advantages over manual bar code based component location tracking, RFID systems are cheaper, fast, more secure and reliable. Another major advantage is that there is no need of line-of-sight for detecting a component in the RFID based location-tracking systems. Hence, even if components are in larger heaps without direct visibility to human eyes, RFID systems can still detect the components using embedded sensors.
  • RFID tags are fitted with the component to be located. These systems contain RFID tags, which are radio frequency transmitters comprising a microchip and antenna and a tiny battery. RFID receiver detects the RF signals emitted by each tag and interprets the encoded identification details.
  • barcode and RFID based technology are not appropriate for locating the position of small components like screws, bolts, nails, pins, etc. as they may not have enough surface area to attach a bar code or a RFID tag.
  • these basic components are used in large quantity and it becomes a tedious task to attach a bar code or RFID tag with each component while archiving in storeroom.
  • the present invention addresses the foregoing need of an inventory management system through which all size and shape of components can tracked before and after an assembly process in a manufacturing facility.
  • the invention address the challenges faced by store keeping clerks in locating the position of excess components returned from assembly floor, which are taken in excess but not being utilized during assembly process. Since basic spare parts which are smaller in size but taken in large quantity from store room due to their frequent usage, excess components needs to be kept back in store keeping units for effective inventory management.
  • the invention recites a method for identifying a component location in an inventory.
  • the first method step is capturing images of the components and then selecting a component from the images of the components.
  • the next method step is extracting features of the selected component and comparing the features of the selected component with the entries of a database. Then displaying a plurality of components from the database wherein at least one feature of the plurality of components is identical to the features of the selected component and selecting another component from the plurality of components that is identical to the previously selected component.
  • FIG. 1 is an illustrative diagram of an environment for the inventory management, in accordance with a described embodiment.
  • FIG. 2 is a diagram depicting a method for managing components in an inventory, in accordance with a described embodiment of the present invention.
  • FIG. 3 is an illustrative schematic diagram of a system for inventory management, in accordance with a described embodiment of the present invention.
  • the present invention relates to methods and systems for inventory management.
  • an inventory management system aids in managing the excess components that are left after an assembly process in a manufacturing unit. It helps a storekeeper to locate the store-keeping unit (SKU) for each component left at the assembly area and updates the database accordingly; thereby the store keeping process becomes easy and highly efficient.
  • SKU store-keeping unit
  • FIG. 1 is an illustrative diagram of an environment for the inventory management, in accordance with a described embodiment.
  • the environment 100 shown in FIG. 1 comprising a handheld device 102 , assembly area 108 and a database unit 106 , in operable communication with the handheld device 102 , is depicted.
  • There are multiple components are present at the assembly area 108 .
  • the components include a nail 110 , a screw 112 , a first bearing 114 , a second bearing 116 and a driller 118 .
  • the components illustrated at assembly area 108 in FIG. 1 is for exemplary purpose only and may include various components, parts, tools, items, assets of various size, shape, weight, or color.
  • the handheld device 102 includes an imaging device 104 , which is capable of capturing still and/or moving images of the assembly area 108 .
  • the handheld device 102 can be a cellular phone, laptop, tablet, or any other mobile device with a camera.
  • the database unit 106 as shown in FIG. 1 is in connection with the handheld device 102 and is capable of sending data to and/or receiving data from the handheld device.
  • the database unit 106 is connected through, either a local area network (LAN), wide area network (WAN), wireless arrangements, or Internet with the handheld device 102 .
  • the database unit 106 contains data of all the available components, parts, tools items, assets etc. in the manufacturing unit.
  • Data stored in the database unit 106 includes the name, image, size, shape, texture, color, weight, available quantity and location of each component, part or tool in a tabular form.
  • the handheld device 102 can retrieve data from the database 106 and is capable of updating the information related to the availability of any component stored in the database unit 106 .
  • FIG. 2 is an illustrative flowchart for managing components in an inventory, in accordance with some embodiments.
  • the handheld device 102 shown in FIG. 1 captures multiple images of the assembly area 108 using the imaging device 104 .
  • the images include all the components left over the assembly area 108 .
  • the handheld device 102 is capable of processing the images and displaying the processed images to a user.
  • the handheld device 102 selects a component, part or tool from the image displayed to the user.
  • the selection of the component could be done based on a predefined selection algorithm or it could be a random selection. A person skilled in the art will appreciate that the user can also do the selection of the component displayed by the handheld device.
  • the handheld device 102 extracts various features of the selected component at the next method step 206 .
  • the extracted component features may include at least one of color, shape, size, weight, texture, dimensions, etc.
  • the component feature extraction could be performed by using various image recognition techniques well known in the art.
  • the system compares extracted features of the selected component with the data stored in the database unit 106 as shown in FIG. 1 . The system can compare either all the features or some of the features of the selected component with the database 106 entries.
  • the handheld device 102 displays a list of components data stored in the database unit 106 that appears to be similar to the selected components based on the comparison performed at step 208 .
  • the component list displayed on the handheld device 102 includes component name, image, color, texture, shape, size, quantity, location etc.
  • the displayed list is in a chronological order i.e. the components having maximum number of features similar to the features extracted from the selected component will appear at the top of the list.
  • the list of the components could also be displayed in alphabetical order, an ascending or descending order based on the available quantity or in any other order preferred by the user.
  • a component is selected from the component list displayed on the handheld device 102 .
  • the component is selected by the handheld device 102 either based on the number of features matched with the features extracted from the component selected previously from the image data or on the bases of a historical event.
  • the historical event is an event stored in the database 106 based on a component selection done by the user previously. The user could also select the component from the component list displayed on the handheld device 102 . If the user does the selection of the component from the displayed list, the selection information gets stored in the database unit 106 as historical event for future reference at step 214 .
  • the user Based on the selection done at method step 212 the user find outs the exact location in a storeroom of the manufacturing unit of the component selected from the image data of the assembly area 106 .
  • the user can easily and efficiently identify the location of all the left overs at the assembly area in the storeroom of the manufacturing unit irrespective to the size, shape, color or texture of the components.
  • FIG. 3 is an illustrative schematic diagram of an inventory management system 300 , in accordance with a described embodiment of the present invention.
  • the inventory management system 300 includes a component search module 302 , component indexing module 304 and a component database 316 .
  • the component search module 302 further includes an imaging unit 306 , an image-preprocessing unit 308 , a feature extraction unit 310 , a feature comparator 312 and location ID unit 314 .
  • the imaging unit 306 is used for capturing images of the area where various components, parts or tools are present such as assembly area 108 shown in FIG. 1 .
  • the imaging unit 306 could be a mobile phone camera, video camera, still camera, webcam, or any other type of image capturing device.
  • the imaging unit 306 is connected to the image-preprocessing unit 308 .
  • the image-preprocessing unit 308 is used to check the quality of the image based on various quality parameters captured by the imaging unit 306 .
  • It quality parameters includes ambient light of the scene, brightness value in the captures image, field of view of the imaging device etc.
  • the image-prepressing unit 308 receives captured image from the imaging unit 306 , it compares the quality parameters values of the captures image with a pre-define minimum quality parameter values stored in the prepossessing unit 308 . If any of the quality parameter value of the captured image is lesser than the pre-stored value of the quality parameter value then the preprocessing unit 308 either rejects the image or prompts an alert message to the user. Once an image captured by the imaging unit 306 passes all the quality parameters checks done by the preprocessing unit 308 , it sends the image to the feature extraction unit 310 .
  • a component of interest is being selected from the components present in the image either using a pre-defined algorithm stored in the component search module or by the user.
  • the method steps followed by the feature extraction unit 310 for the selection of the component of interest are same as step 204 and step 206 of FIG. 2 explained earlier.
  • Another function of the feature extraction unit 310 is to extract features of the component of the interest.
  • the features extracted from the component may include color, shape, texture, size, etc.
  • the extracted features of the component send to the feature comparator 312 as shown in FIG. 3 .
  • the comparator 312 compares the extracted features with the data entries present in the component database 316 .
  • the component database 316 contains data of all the available components, parts, tools etc. in the manufacturing unit.
  • Data entries in the component database 316 includes the name, image, size, shape, texture, color, weight, available quantity and location ID of each component, part or tool.
  • the location ID of a component is the location in the storeroom of the manufacturing unit where a similar kind of components is stored.
  • the comparator 312 extract a list of components from the database 316 based on the comparison of the extracted features of the component in interest with database entries. The extracted list is displayed on the handheld device 102 as shown in FIG. 1 . Then the location ID unit 314 selects a component that is same as the component of interest from the component list extracted by the feature comparator 312 by following the method recited earlier at step 212 of FIG. 2 .
  • the selection is considered, as historical event and it get stored in the database 316 for future reference.
  • the location ID associated with the component selected by the location ID unit 314 from the extracted list is presented to the user as the location information of the component of interest.
  • the function of the component indexing module 304 is to create and update the database 316 that can in turn used for locating the location ID of any component by the component search module 302 .
  • the indexing module 304 includes an imaging unit 318 , an image-processing unit 320 , a component feature extraction unit 322 and a feature-indexing unit 324 .
  • the imaging unit 318 is similar to the imaging unit 306 and is used for capturing images of the individual components, parts and tools available in the storeroom of the manufacturing unit. Component images should be capture in the predetermined background, with as many angles as possible and with the best camera settings captured by the imaging device 318 for good and precise visibility of the components.
  • the images of the individual components, parts and tools are sent image-processing unit 320 for further processing.
  • the image-processing unit 320 removes noise and enhances the image quality using image-processing techniques known in the art.
  • the processed images from processing unit 320 are sent to the component feature extraction unit 322 .
  • the extraction unit 322 extracts the features of the component present in the image received from the processing unit 320 .
  • the features extracted by the extraction unit 322 include color, texture, size, and shape.
  • the next functional block is feature-indexing unit 324 receives the extracted features from the extraction unit 322 and store the feature details along with the image of each component in the component database 316 .
  • a user can enter the component name, available quantity in the storeroom and the location ID information for each component stored by the indexing unit 322 in the database 316 .
  • a means for inputting component name, available quantity and the location ID information user input means are provided with the database 316 .
  • the user can also input one or more comments related to the component present in the image.
  • the comments may include an observation stating a defect in the component, a command related to use of the component, a new type of component entry in the existing database entries and so on.
  • the user input means can also be provided along with the indexing unit 322 and the user can add information of each component along with the extracted features and component images that in turns stored as an individual component file in the database 316 by the indexing unit 322 .
  • the component indexing module 304 create and maintain the component database 316 for the component search module 302 by storing information of the available and newly arrived components in the storeroom.
  • Embodiments of the present invention enable faster, easier, more efficient and less tiring usage of inventory management.
  • the present invention is applicable for all types, shape, and size of components; hence, it eliminates the size dependency limitation from the inventory management system.
  • a processor base system such as general purpose or special purpose computer.
  • different implementations of the present technique may perform some or all the steps described herein in different orders or substantially concurrently, that is, in parallel.
  • the functions may be implemented in a variety of programming languages.
  • Such code as will be appreciated by those of ordinary skilled in the art, may be stored or adapted for storage in one or more tangible machine readable media, such as on memory chips, local or remote hard disks, optical disks or other media, which may be accessed by a processor based system to execute the stored code.

Abstract

Embodiments of the present invention relate to an inventory management method and system. The present invention addresses the foregoing need of an inventory management system through which all size and shape of components can tracked before and after an assembly process in a manufacturing facility. The invention address the challenges faced by store keeping clerks in locating the position of excess components returned from assembly floor, which are taken in excess but not being utilized during assembly process. Since basic spare parts which are smaller in size but taken in large quantity from store room due to their frequent usage, excess components needs to be kept back in store keeping units for effective inventory management.

Description

    RELATED APPLICATION DATA
  • This application claims priority to India Patent Application No. 2946/CHE/2012, filed Jul. 18, 2012, the disclosure of which is hereby incorporated by reference in its entirety.
  • FIELD OF THE INVENTION
  • The invention relates generally to a method and system for inventory management. In particular, it relates to a method and system for locating the position of components in store keeping unit for computer assisted inventory placement and auditing in manufacturing plant.
  • BACKGROUND
  • To run a manufacturing unit successfully, it is important for companies to have a detailed, up-to-date record of their inventory. With the advent of computers, inputting and tracking inventories become easier by inventory management software and systems. For effective inventory management major technologies used are bar codes and RFID. One known bar code technique uses unique bar code patterns for each component and these codes are mapped to component details such as make, model number, location details in a computer system. Using a bar code reader, a storeroom clerk captures the barcode of the component and the details of that component are retrieved from the system.
  • RFID technology is another alternate technology widely used for component and asset tracking in industries. RFID based solutions have several advantages over manual bar code based component location tracking, RFID systems are cheaper, fast, more secure and reliable. Another major advantage is that there is no need of line-of-sight for detecting a component in the RFID based location-tracking systems. Hence, even if components are in larger heaps without direct visibility to human eyes, RFID systems can still detect the components using embedded sensors. In this system, RFID tags are fitted with the component to be located. These systems contain RFID tags, which are radio frequency transmitters comprising a microchip and antenna and a tiny battery. RFID receiver detects the RF signals emitted by each tag and interprets the encoded identification details.
  • One major limitation of above barcode and RFID based technology is that they are not appropriate for locating the position of small components like screws, bolts, nails, pins, etc. as they may not have enough surface area to attach a bar code or a RFID tag. In addition to this, these basic components are used in large quantity and it becomes a tedious task to attach a bar code or RFID tag with each component while archiving in storeroom.
  • One solution to solve above problem is by collecting all smaller items in one container and attaching the bar code to container, but this also time consuming and may not help in identifying the components once brought outside the container.
  • SUMMARY
  • The present invention addresses the foregoing need of an inventory management system through which all size and shape of components can tracked before and after an assembly process in a manufacturing facility.
  • The invention address the challenges faced by store keeping clerks in locating the position of excess components returned from assembly floor, which are taken in excess but not being utilized during assembly process. Since basic spare parts which are smaller in size but taken in large quantity from store room due to their frequent usage, excess components needs to be kept back in store keeping units for effective inventory management.
  • The invention recites a method for identifying a component location in an inventory. The first method step is capturing images of the components and then selecting a component from the images of the components. The next method step is extracting features of the selected component and comparing the features of the selected component with the entries of a database. Then displaying a plurality of components from the database wherein at least one feature of the plurality of components is identical to the features of the selected component and selecting another component from the plurality of components that is identical to the previously selected component.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • These and other features, aspects, and advantages of the present invention will be better understood when the following detailed description is read with reference to the accompanying drawings in which like characters represent like parts throughout the drawings, wherein:
  • FIG. 1 is an illustrative diagram of an environment for the inventory management, in accordance with a described embodiment.
  • FIG. 2 is a diagram depicting a method for managing components in an inventory, in accordance with a described embodiment of the present invention.
  • FIG. 3 is an illustrative schematic diagram of a system for inventory management, in accordance with a described embodiment of the present invention.
  • While systems and methods are described herein by way of example and embodiments, those skilled in the art recognize that systems and methods for inventory management are not limited to the embodiments or drawings described. It should be understood that the drawings and description are not intended to be limiting to the particular form disclosed. Rather, the intention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the appended claims. Any headings used herein are for organizational purposes only and are not meant to limit the scope of the description or the claims. As used herein, the word “may” is used in a permissive sense (i.e., meaning having the potential to) rather than the mandatory sense (i.e., meaning must). Similarly, the words “include”, “including”, and “includes” mean including, but not limited to.
  • DETAILED DESCRIPTION
  • The following description is the full and informative description of the best method and system presently contemplated for carrying out the present invention which is known to the inventors at the time of filing the patent application.
  • The present invention relates to methods and systems for inventory management.
  • According to one preferred embodiment, an inventory management system is provided. This embodiment aids in managing the excess components that are left after an assembly process in a manufacturing unit. It helps a storekeeper to locate the store-keeping unit (SKU) for each component left at the assembly area and updates the database accordingly; thereby the store keeping process becomes easy and highly efficient.
  • FIG. 1 is an illustrative diagram of an environment for the inventory management, in accordance with a described embodiment. The environment 100 shown in FIG. 1 comprising a handheld device 102, assembly area 108 and a database unit 106, in operable communication with the handheld device 102, is depicted. There are multiple components are present at the assembly area 108. The components include a nail 110, a screw 112, a first bearing 114, a second bearing 116 and a driller 118. It is apparent to those of ordinary skill in the art that the components illustrated at assembly area 108 in FIG. 1 is for exemplary purpose only and may include various components, parts, tools, items, assets of various size, shape, weight, or color.
  • The handheld device 102 includes an imaging device 104, which is capable of capturing still and/or moving images of the assembly area 108. The handheld device 102 can be a cellular phone, laptop, tablet, or any other mobile device with a camera. The database unit 106 as shown in FIG. 1 is in connection with the handheld device 102 and is capable of sending data to and/or receiving data from the handheld device. The database unit 106 is connected through, either a local area network (LAN), wide area network (WAN), wireless arrangements, or Internet with the handheld device 102. The database unit 106 contains data of all the available components, parts, tools items, assets etc. in the manufacturing unit. Data stored in the database unit 106 includes the name, image, size, shape, texture, color, weight, available quantity and location of each component, part or tool in a tabular form. The handheld device 102 can retrieve data from the database 106 and is capable of updating the information related to the availability of any component stored in the database unit 106.
  • FIG. 2 is an illustrative flowchart for managing components in an inventory, in accordance with some embodiments. In the manufacturing unit after the assembly process completed some left overs such as components, tools, parts, items, assets etc. remains at the product assembly area 108 as shown in FIG. 1. At the first step 202, the handheld device 102 shown in FIG. 1 captures multiple images of the assembly area 108 using the imaging device 104. The images include all the components left over the assembly area 108. The handheld device 102 is capable of processing the images and displaying the processed images to a user. At step 204, the handheld device 102 selects a component, part or tool from the image displayed to the user. The selection of the component could be done based on a predefined selection algorithm or it could be a random selection. A person skilled in the art will appreciate that the user can also do the selection of the component displayed by the handheld device. The handheld device 102 extracts various features of the selected component at the next method step 206. The extracted component features may include at least one of color, shape, size, weight, texture, dimensions, etc. The component feature extraction could be performed by using various image recognition techniques well known in the art. At step 208, the system compares extracted features of the selected component with the data stored in the database unit 106 as shown in FIG. 1. The system can compare either all the features or some of the features of the selected component with the database 106 entries. A person ordinarily skilled in the art will appreciate that the user can select the extracted feature of the selected component for the comparison with the database entries. At the next step 210, the handheld device 102 displays a list of components data stored in the database unit 106 that appears to be similar to the selected components based on the comparison performed at step 208. The component list displayed on the handheld device 102 includes component name, image, color, texture, shape, size, quantity, location etc. The displayed list is in a chronological order i.e. the components having maximum number of features similar to the features extracted from the selected component will appear at the top of the list. The list of the components could also be displayed in alphabetical order, an ascending or descending order based on the available quantity or in any other order preferred by the user. At method step 212, a component is selected from the component list displayed on the handheld device 102. The component is selected by the handheld device 102 either based on the number of features matched with the features extracted from the component selected previously from the image data or on the bases of a historical event. The historical event is an event stored in the database 106 based on a component selection done by the user previously. The user could also select the component from the component list displayed on the handheld device 102. If the user does the selection of the component from the displayed list, the selection information gets stored in the database unit 106 as historical event for future reference at step 214. Based on the selection done at method step 212 the user find outs the exact location in a storeroom of the manufacturing unit of the component selected from the image data of the assembly area 106. By following the method steps show in FIG. 2 and described above the user can easily and efficiently identify the location of all the left overs at the assembly area in the storeroom of the manufacturing unit irrespective to the size, shape, color or texture of the components.
  • FIG. 3 is an illustrative schematic diagram of an inventory management system 300, in accordance with a described embodiment of the present invention. The inventory management system 300 includes a component search module 302, component indexing module 304 and a component database 316.
  • The component search module 302 further includes an imaging unit 306, an image-preprocessing unit 308, a feature extraction unit 310, a feature comparator 312 and location ID unit 314. The imaging unit 306 is used for capturing images of the area where various components, parts or tools are present such as assembly area 108 shown in FIG. 1. The imaging unit 306 could be a mobile phone camera, video camera, still camera, webcam, or any other type of image capturing device. The imaging unit 306 is connected to the image-preprocessing unit 308. The image-preprocessing unit 308 is used to check the quality of the image based on various quality parameters captured by the imaging unit 306. It quality parameters includes ambient light of the scene, brightness value in the captures image, field of view of the imaging device etc. when the image-prepressing unit 308 receives captured image from the imaging unit 306, it compares the quality parameters values of the captures image with a pre-define minimum quality parameter values stored in the prepossessing unit 308. If any of the quality parameter value of the captured image is lesser than the pre-stored value of the quality parameter value then the preprocessing unit 308 either rejects the image or prompts an alert message to the user. Once an image captured by the imaging unit 306 passes all the quality parameters checks done by the preprocessing unit 308, it sends the image to the feature extraction unit 310. In the feature extraction unit 310, a component of interest is being selected from the components present in the image either using a pre-defined algorithm stored in the component search module or by the user. The method steps followed by the feature extraction unit 310 for the selection of the component of interest are same as step 204 and step 206 of FIG. 2 explained earlier. Another function of the feature extraction unit 310 is to extract features of the component of the interest. The features extracted from the component may include color, shape, texture, size, etc. The extracted features of the component send to the feature comparator 312 as shown in FIG. 3. The comparator 312 compares the extracted features with the data entries present in the component database 316. The component database 316 contains data of all the available components, parts, tools etc. in the manufacturing unit. Data entries in the component database 316 includes the name, image, size, shape, texture, color, weight, available quantity and location ID of each component, part or tool. The location ID of a component is the location in the storeroom of the manufacturing unit where a similar kind of components is stored. The comparator 312 extract a list of components from the database 316 based on the comparison of the extracted features of the component in interest with database entries. The extracted list is displayed on the handheld device 102 as shown in FIG. 1. Then the location ID unit 314 selects a component that is same as the component of interest from the component list extracted by the feature comparator 312 by following the method recited earlier at step 212 of FIG. 2. If the user does the selection of the component from the extracted list then the selection is considered, as historical event and it get stored in the database 316 for future reference. The location ID associated with the component selected by the location ID unit 314 from the extracted list is presented to the user as the location information of the component of interest.
  • Function of the component indexing module 304 is to create and update the database 316 that can in turn used for locating the location ID of any component by the component search module 302. The indexing module 304 includes an imaging unit 318, an image-processing unit 320, a component feature extraction unit 322 and a feature-indexing unit 324. The imaging unit 318 is similar to the imaging unit 306 and is used for capturing images of the individual components, parts and tools available in the storeroom of the manufacturing unit. Component images should be capture in the predetermined background, with as many angles as possible and with the best camera settings captured by the imaging device 318 for good and precise visibility of the components. The images of the individual components, parts and tools are sent image-processing unit 320 for further processing. The image-processing unit 320 removes noise and enhances the image quality using image-processing techniques known in the art. The processed images from processing unit 320 are sent to the component feature extraction unit 322. The extraction unit 322 extracts the features of the component present in the image received from the processing unit 320. The features extracted by the extraction unit 322 include color, texture, size, and shape. The next functional block is feature-indexing unit 324 receives the extracted features from the extraction unit 322 and store the feature details along with the image of each component in the component database 316. A user can enter the component name, available quantity in the storeroom and the location ID information for each component stored by the indexing unit 322 in the database 316. A means for inputting component name, available quantity and the location ID information user input means are provided with the database 316. The user can also input one or more comments related to the component present in the image. The comments may include an observation stating a defect in the component, a command related to use of the component, a new type of component entry in the existing database entries and so on. A person skilled in the art will appreciate that the user input means can also be provided along with the indexing unit 322 and the user can add information of each component along with the extracted features and component images that in turns stored as an individual component file in the database 316 by the indexing unit 322. As recited, the component indexing module 304 create and maintain the component database 316 for the component search module 302 by storing information of the available and newly arrived components in the storeroom.
  • Embodiments of the present invention enable faster, easier, more efficient and less tiring usage of inventory management. The present invention is applicable for all types, shape, and size of components; hence, it eliminates the size dependency limitation from the inventory management system.
  • The present description includes the best presently contemplated method for carrying out the present invention. Various modifications to the embodiments will be readily apparent to those skilled in the art and some features of the present invention may be used without the corresponding use of other features. Accordingly, the present invention is not intended to be limited to the embodiments shown but is to be accorded the widest scope consistent with the principles and features described herein.
  • As will be appreciated by those ordinary skilled in the art, the aforementioned example, demonstrations, and method steps may be implemented by suitable code on a processor base system, such as general purpose or special purpose computer. It should also be noted that different implementations of the present technique may perform some or all the steps described herein in different orders or substantially concurrently, that is, in parallel. Furthermore, the functions may be implemented in a variety of programming languages. Such code, as will be appreciated by those of ordinary skilled in the art, may be stored or adapted for storage in one or more tangible machine readable media, such as on memory chips, local or remote hard disks, optical disks or other media, which may be accessed by a processor based system to execute the stored code.

Claims (33)

What is claimed is:
1. A method for identifying a component location in an inventory, the method comprising:
capturing at least one image of one or more components;
selecting a first component from the at least one image of one or more components;
extracting one or more features of the first component;
comparing the one or more features of the first component with one or more entries of a database;
displaying a plurality of components from the database wherein at least one feature of the plurality of components is identical to the one or more features of the first component; and
selecting a second component from the plurality of components wherein the second component is identical to the first component.
2. The method as claimed in claim 1 further comprising prompting an alert message before capturing at least one image of the one or more components if a value of one or more image quality parameters is below a predefined value of the one or more image quality parameters.
3. The method as claimed in claim 2 wherein the one or more image quality parameters comprise at least one of ambient light, brightness and field of view of an image-capturing device.
4. The method as claimed in claim 1 wherein selecting the first component from at least one image comprises selecting the component based on a user input.
5. The method as claimed in claim 1 wherein selecting the first component from the at least one image comprises selecting the component using one or more selection algorithms.
6. The method as claimed in claim 1 wherein extracting the one or more features of the first component comprises extracting the one or more features using one or more image recognition techniques.
7. The method as claimed in claim 1 wherein the one or more features comprise one or more of a size, a shape, a color and a texture of the first component.
8. The method as claimed in claim 1 wherein comparing the one or more features of the first component comprises selecting at least one feature of the first component for comparison.
9. The method as claimed in claim 8 wherein selecting at least one feature of the first component comprises selecting at least one feature based on a user input.
10. The method as claimed in claim 8 wherein selecting at least one feature of the first component comprises selecting at least one feature based on a historical event.
11. The method as claimed in claim 1 wherein the entries of the database comprise one or more of name, image, location information, type, color, size, texture, shape, unit ID, quantity and use of a plurality of components.
12. The method as claimed in claim 1 wherein displaying a list of the plurality of components further comprises displaying one or more of name, image, location information, type, color, size, texture, shape, unit ID, quantity and use of the plurality of components.
13. The method as claimed in claim 1 wherein selecting the second component comprises selecting the second component based on a user input.
14. The method as claimed in claim 1 wherein selecting the second component comprises selecting the second component based on a historical event.
15. The method as claimed in claim 1 wherein selecting the second component further comprises storing the second component selection as a historical event.
16. A system for identifying a component location in an inventory, the system comprising:
a handheld device for capturing at least one image of one or more components;
a first selection module for selecting a first component from the at least one image of the one or more components;
extracting one or more features of the first component by a processor;
a comparison module for comparing the one or more features of the first component with one or more entries of a database wherein the database is linked with the handheld device;
a display for displaying a plurality of components from the database, wherein at least one feature of the plurality of components displayed is identical to the one or more features of the first component; and
a second selection module for selecting a second component from the plurality of components, wherein the second component is identical to the first component.
17. The system as claimed in claim 16 further comprising an alarm for prompting an alert message before capturing at least one image of the one or more components if a value of one or more image quality parameters is below a predefined value of the one or more image quality parameters.
18. The system as claimed in claim 16 wherein the first selection module further comprises a user input module for selecting the first component from the at least one image of the one or more components.
19. The system as claimed in claim 16 wherein the first selection module further comprises a selection module for executing one or more selection algorithms for selecting the first component from the at least one image of the one or more components.
20. The system as claimed in claim 16 further comprises an image recognition module.
21. The system as claimed in claim 16 wherein the comparison module for comparing the one or more features of the first component comprises a third selection module for selecting at least one feature of the first component for comparison.
22. The system as claimed in claim 21 wherein the third selection module further comprises a user input module for selecting at least one feature of the first component.
23. The system as claimed in claim 16 wherein the second selection module comprises a user input module for selecting the second component.
24. The system as claimed in claim 16 wherein the second selection module comprises a memory element for storing the second component selection as a historical event.
25. A computer program product for use with a computer, the computer program product comprising a computer usable medium having a computer readable program code embodied therein system for identifying a component location in an inventory, the computer readable code performing:
capturing at least one image of one or more components;
selecting a first component from the at least one image of one or more components;
extracting one or more features of the first component;
comparing the one or more features of the first component with one or more entries of a database;
displaying a plurality of components from the database wherein at least one feature of the plurality of components is identical to the one or more features of the first component; and
selecting a second component from the plurality of components wherein the second component is identical to the first component.
26. The computer program product as claimed in claim 25 further performing prompting an alert message before capturing at least one image of the one or more components if a value of one or more image quality parameters is below a predefined value of the one or more image quality parameters.
27. The computer program product as claimed in claim 25 wherein selecting the first component comprises receiving a user input for selecting the first component from the at least one image of the one or more components.
28. The computer program product as claimed in claim 25 wherein selecting the first component from the at least one image by executing one or more selection algorithms.
29. The computer program product as claimed in claim 25 wherein extracting the one or more features of the first component using one or more image recognition techniques.
30. The computer program product as claimed in claim 25 wherein comparing the one or more features of the first component comprises selecting at least one feature of the first component for comparison.
31. The computer program product as claimed in claim 30 wherein selecting at least one feature of the first component for comparison further comprises receiving a user input for selecting at least one feature of the first component.
32. The computer program product as claimed in claim 25 wherein selecting a second component from the plurality of components comprises receiving a user input for selecting the second component.
33. The computer program product as claimed in claim 25 wherein selecting a second component from the plurality of components comprises storing the second component selection as a historical event.
US13/945,403 2012-07-19 2013-07-18 Methods and systems for enabling vision based inventory management Active 2035-05-14 US9665794B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
IN2946CH2012 2012-07-19
IN2946/CHE/2012 2012-07-19

Publications (3)

Publication Number Publication Date
US20140023276A1 US20140023276A1 (en) 2014-01-23
US20160224865A9 true US20160224865A9 (en) 2016-08-04
US9665794B2 US9665794B2 (en) 2017-05-30

Family

ID=49946591

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/945,403 Active 2035-05-14 US9665794B2 (en) 2012-07-19 2013-07-18 Methods and systems for enabling vision based inventory management

Country Status (1)

Country Link
US (1) US9665794B2 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109492968A (en) * 2018-10-29 2019-03-19 西北工业大学 Cutter treasury management system based on database technology
US11599742B2 (en) * 2020-04-22 2023-03-07 Dell Products L.P. Dynamic image recognition and training using data center resources and data

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070124216A1 (en) * 2000-03-07 2007-05-31 Michael Lucas Systems and methods for locating and purchasing proximal inventory items
EP1436742A1 (en) 2001-09-18 2004-07-14 Pro-Corp Holdings International Limited Image recognition inventory management system
US20050137943A1 (en) 2003-12-17 2005-06-23 Ncr Corporation Method and system for assisting a search for articles within a storage facility
US7657100B2 (en) * 2005-05-09 2010-02-02 Like.Com System and method for enabling image recognition and searching of images
US20100092093A1 (en) * 2007-02-13 2010-04-15 Olympus Corporation Feature matching method
US9147174B2 (en) 2008-08-08 2015-09-29 Snap-On Incorporated Image-based inventory control system using advanced image recognition
US8243166B2 (en) * 2009-01-20 2012-08-14 Lockheed Martin Corporation Automatic detection of blocked field-of-view in camera systems
US8908962B2 (en) * 2011-09-30 2014-12-09 Ebay Inc. Item recommendations using image feature data

Also Published As

Publication number Publication date
US9665794B2 (en) 2017-05-30
US20140023276A1 (en) 2014-01-23

Similar Documents

Publication Publication Date Title
US9697429B2 (en) Method and apparatus for image processing to avoid counting shelf edge promotional labels when counting product labels
US10078812B2 (en) Data center infrastructure management system having real time enhanced reality tablet
US9311799B2 (en) Modifying RFID system operation using movement detection
US10296814B1 (en) Automated and periodic updating of item images data store
JP2019055828A (en) Shelf information estimation device and information processing program
CN104462530A (en) Method and device for analyzing user preferences and electronic equipment
WO2011143748A1 (en) System and method for tracking items
US20140244392A1 (en) Graphical Recognition Inventory Management and Marketing System
US20160098427A1 (en) Filtering inventory objects using images in an rfid system
US20200193281A1 (en) Method for automating supervisory signal during training of a neural network using barcode scan
US8126198B2 (en) Method for auditing and maintaining an ordered inventory
US20200192608A1 (en) Method for improving the accuracy of a convolution neural network training image data set for loss prevention applications
CN111159452A (en) Method and system for reminding forgetful article, computer equipment and storage medium
US9147109B2 (en) Method for adding business card information into contact list
JP6687199B2 (en) Product shelf position registration program and information processing device
US10402777B2 (en) Method and a system for object recognition
US9665794B2 (en) Methods and systems for enabling vision based inventory management
CN109213397B (en) Data processing method and device and user side
US11080977B2 (en) Management system, server, management device, and management method
US20210209550A1 (en) Systems, apparatuses, and methods for triggering object recognition and planogram generation via shelf sensors
CN113902398A (en) Asset inventory system, method, computer device and storage medium
CN109635688B (en) Method and system for managing books on bookshelf based on image recognition
WO2021138451A1 (en) A system and method for tracking wine in a wine-cellar and monitoring inventory
Merrad et al. A Real-time Mobile Notification System for Inventory Stock out Detection using SIFT and RANSAC.
Friedel et al. Asset identification using image descriptors

Legal Events

Date Code Title Description
AS Assignment

Owner name: INFOSYS LIMITED, INDIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:RAI, HARIKRISHNA GANDHINAGARA NARAYANA;DEEPAK, KRISHNAMURTHY SAI;PISIPATI, RADHA KRISHNA;SIGNING DATES FROM 20131211 TO 20140516;REEL/FRAME:033341/0408

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4