WO2023138447A1 - Ai weighing system, and method for improving precision of ai model by using various types of data sets - Google Patents

Ai weighing system, and method for improving precision of ai model by using various types of data sets Download PDF

Info

Publication number
WO2023138447A1
WO2023138447A1 PCT/CN2023/071665 CN2023071665W WO2023138447A1 WO 2023138447 A1 WO2023138447 A1 WO 2023138447A1 CN 2023071665 W CN2023071665 W CN 2023071665W WO 2023138447 A1 WO2023138447 A1 WO 2023138447A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
data
processing
weighing system
model
Prior art date
Application number
PCT/CN2023/071665
Other languages
French (fr)
Chinese (zh)
Inventor
朱辰
西田翔
杉山聪
Original Assignee
索尼半导体解决方案公司
索尼集团公司
朱辰
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 索尼半导体解决方案公司, 索尼集团公司, 朱辰 filed Critical 索尼半导体解决方案公司
Publication of WO2023138447A1 publication Critical patent/WO2023138447A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01GWEIGHING
    • G01G19/00Weighing apparatus or methods adapted for special purposes not provided for in the preceding groups
    • G01G19/40Weighing apparatus or methods adapted for special purposes not provided for in the preceding groups with provisions for indicating, recording, or computing price or other quantities dependent on the weight
    • G01G19/413Weighing apparatus or methods adapted for special purposes not provided for in the preceding groups with provisions for indicating, recording, or computing price or other quantities dependent on the weight using electromechanical or electronic computing means
    • G01G19/414Weighing apparatus or methods adapted for special purposes not provided for in the preceding groups with provisions for indicating, recording, or computing price or other quantities dependent on the weight using electromechanical or electronic computing means using electronic computing means only
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes

Definitions

  • the present invention relates to an AI (Antificial Intelligence) weighing system, more specifically, relates to an AI weighing (AI scale) system utilizing an image sensor with an AI processing function.
  • AI Artificial Intelligence
  • AI artificial intelligence
  • Such an AI weighing device automatically recognizes a product by establishing an intelligent recognition model and analyzing a product image using the model.
  • Such an AI weighing device avoids the steps of manual weighing and printing barcodes, and avoids the need to manually search or memorize the prices of various commodities, thereby greatly improving the work efficiency of places such as supermarkets and saving labor costs.
  • the identification of goods is carried out by sending the images captured by the camera to an external computing device such as the cloud, and the learning, establishment, training and retraining of the AI model are also carried out on the cloud.
  • data transmission depends on data traffic, which may be unstable and have delays or even errors, which will significantly affect the recognition speed and accuracy.
  • cloud AI recognition depends on the stability of the network and cannot be implemented offline. Therefore, cloud AI recognition significantly depends on the cloud environment.
  • the pictures sent to the cloud may contain private information, such as the user's personal information, which poses a risk of privacy leakage.
  • the recognition accuracy of the recognition model of the existing AI weighing device needs to be improved.
  • a place such as a supermarket
  • various packages, sales forms, and the like For example, for fresh goods, there are various packaging forms such as boxed and plastic bags, and may also be packaged by itself; at the same time, they may be sold whole or divided (for example, sliced, divided into pieces, etc.). All of these make it difficult to accurately identify products.
  • the present invention can perform AI recognition offline by using an image sensor with AI processing function, thereby significantly improving the recognition accuracy and recognition speed of the intelligent weighing device, and can be used offline.
  • the present invention also significantly improves the recognition accuracy of the AI model by utilizing image data containing various data sets.
  • the AI weighing system may include: a storage platform for placing a target object, the storage platform can weigh the target object; a first camera device for photographing and identifying the target object placed on the storage platform, the first camera device includes an image sensor, and the image sensor can perform AI processing for identifying the target object in an offline state; and an output unit is used for outputting the recognition result and the weighing result of the target object.
  • the image sensor of the camera device of the AI weighing system may be a CMOS image sensor chip, the CMOS image sensor chip includes a first substrate and a second substrate, the first substrate has a plurality of pixels for converting optical signals into electrical signals, and the second substrate has a memory and a processing circuit, the memory stores an AI model, and the processing circuit has a function of performing the AI processing based on the electrical signal by using the AI model.
  • the AI models stored on the image sensor include a first inference model.
  • the processing circuit of the stacked CMOS image sensor chip generates image data
  • the processing circuit may include: a learning part that retrains the AI model based on the image data; and an inference part that uses the AI model to recognize the target object.
  • the AI weighing system may further include one or more computing devices located in a cloud environment, and the one or more computing devices have corresponding processors and memories.
  • the processing circuit of the stacked CMOS image sensor chip of the first camera generates image data, and the image data is sent to the one or more computing devices.
  • the one or more computing devices create a second inference model based on the image data generated by the stacked CMOS image sensor chip, and may directly deploy the second inference model into the memory of the stacked CMOS image sensor chip such that the first inference model is updated.
  • the stacked CMOS image sensor chip is capable of selecting a size of the image data including a full sensor size and a Video Graphics Array (VGA) size based on the AI processing of the processing circuit.
  • VGA Video Graphics Array
  • the AI processing when the stacked CMOS image sensor outputs the image data having a full sensor size, the AI processing includes intercepting an overall image or a partial image of the target object from the image data having the full sensor size, and the partial image includes a VGA-sized image.
  • the image data captured by the first camera device of the AI weighing system may include: a profile image of the target object; and/or a cross-sectional image of the target object; and/or a packaging image of the target object.
  • the image data captured by the first camera device of the AI weighing system may also include a partially enlarged image of the target object.
  • the first camera device of the AI weighing system may further include a ToF sensor, and the AI processing includes combining RBG data output by the image sensor and ToF data output by the ToF sensor.
  • the first camera device of the AI weighing system may further include a multi-wavelength sensor, and the AI processing can utilize the output of the multi-wavelength sensor to optimize the identification of the target object.
  • the first camera device of the AI weighing system may further include a polarized light sensor, and the AI processing can utilize the output of the polarized light sensor to optimize the identification of the target object.
  • the AI weighing system may further include: a second camera device for capturing an image of the target person and sending it to the first camera device, and wherein the AI processing includes acquiring feature data from the image of the target person captured by the second camera device and using the feature data to assist in identifying the target object.
  • the feature data includes anonymous feature data such as gender and age of the target person.
  • the AI weighing system may further include: a third camera device, configured to acquire an image of the target person and send it to the first camera device, and wherein the AI processing includes performing SLAM processing on the image of the target person captured by the third camera device, and outputting metadata of the processing result to assist in identifying the target object.
  • the third camera device also acquires an image of the shopping cart or shopping basket of the target person and sends it to the first camera device, and the AI processing includes performing SLAM processing on the image of the shopping cart or shopping basket, and outputting metadata of the processing result to assist in identifying the target object.
  • the third camera device of the present invention may include a plurality of image sensors positioned at different positions within the moving range of the target person.
  • the AI processing of the AI weighing system may also include obtaining other information including ambient temperature, area address and/or weather conditions, and assist in identifying the target object based on the other information.
  • the method for increasing the recognition accuracy of an AI model by using multiple data sets includes: acquiring image data of an item; creating an AI model using learning data, the learning data including the image data and the name and attribute of the item; applying the AI model to identify the target object, wherein the learning data includes at least two of the following three data sets: an outline image of the item; a cross-sectional image of the item; and a packaging image of the item.
  • the learning data used to create the AI model may also include a data set of partially enlarged images of the item.
  • the image data of the item includes RGB image data and at least one of the following data: ToF data, multi-wavelength data, polarization data.
  • the recognition speed and precision of the AI weighing system are significantly improved, and especially, the target object can be recognized accurately and quickly even in an offline state. Also, the method for creating an AI model according to the present invention significantly improves the accuracy of the recognition model by adding different data sets.
  • FIG. 1 is a schematic diagram showing a first embodiment of an AI weighing system according to the present invention.
  • FIG. 2 is a schematic diagram showing an embodiment of an image sensor according to the present invention.
  • FIG. 3 is a schematic diagram showing a stacked CMOS image sensor chip according to the present invention.
  • FIG. 4 is a flowchart illustrating steps for creating, training or retraining an AI model according to the present invention.
  • FIG. 5 is a flowchart illustrating an AI recognition method according to the present invention.
  • FIG. 6 is a schematic diagram illustrating an image output and/or processing mode of an image sensor according to the present invention.
  • Fig. 7 is a schematic diagram showing a second embodiment of the AI weighing system according to the present invention.
  • Fig. 8 is a schematic diagram showing a third embodiment of the AI weighing system according to the present invention.
  • Fig. 9 is a schematic diagram showing a fourth embodiment of the AI weighing system according to the present invention.
  • FIG. 10 is a schematic diagram showing a fifth embodiment of the AI weighing system according to the present invention.
  • the AI weighing system includes: a storage platform 1 , a camera device 2 and an output unit 3 , wherein an object 4 is placed on the storage platform 1 .
  • the storage table 1 has a built-in weighing device, which can weigh the target object 4 .
  • the output unit 3 can be a display as shown in FIG. 1 , such as a liquid crystal display, to display the recognition result and weighing result of the target object 4 to the user.
  • the output unit 3 may also output the recognition result and the weighing result in audio form.
  • the output unit 3 can be a touch display to interact with the user.
  • the object of the present invention may also be goods stored in a warehouse.
  • the imaging device 2 is, for example, arranged above the object table 1 to photograph the target object.
  • the imaging device 2 includes an image sensor 21 such as a CMOS image sensor.
  • the image sensor 21 according to the present invention is capable of not only storing and outputting captured images, but also performing various processing including AI processing on image data.
  • AI processing includes obtaining various information (metadata such as feature data of objects) from image data, and recognizing objects in images.
  • the image sensor 21 can also communicate with a cloud environment (cloud).
  • cloud cloud
  • the AI weighing system of the present invention does not depend on network connections and cloud servers, and can quickly and accurately identify objects even in an offline state.
  • FIG. 2 shows a functional block diagram of an image sensor 21 according to the invention.
  • the image sensor 21 includes an imaging section 211 , a memory 212 and an AI processing section 213 .
  • the image sensor 21 further includes a control unit (not shown) that controls the imaging unit.
  • the imaging unit 211 photographs the target object, sends the imaging data to the AI processing unit 213 , and stores the imaging data in the memory 212 .
  • the memory 212 stores an AI model, for example, an inference model for recognizing a target ("first inference model").
  • the AI model is, for example, a neural network computing model for computer vision created using learning of a deep neural network (DNN) by executing a program, for example, stored in memory 212 and/or in a cloud environment.
  • the AI model may be a learning model utilizing a multi-layer neural network. It should be understood that for the AI model here, any suitable known AI model and algorithm may be selected according to usage and requirements.
  • the AI processing unit 213 is, for example, a graphics processing unit (Graphic Processing Unit, GPU), so that the image data of the target object can be processed using the AI inference model stored in the memory 212, and the processing result is sent to the output unit 3.
  • a graphics processing unit Graphic Processing Unit, GPU
  • the AI processing unit 213 may include an inference unit 2131 and a learning unit 2132 .
  • the inference unit 2131 uses the AI inference model stored in the memory 212 to recognize the target image captured and sent by the imaging unit 211 , and sends the recognition result to the output unit 3 .
  • the learning unit 2132 retrains the AI inference model based on the object image and the confirmed recognition result.
  • the learning unit 2132 can also use data in the cloud environment to retrain the AI inference model. Retraining can be done when the AI weighing system is idle and/or optionally when the network is available, so as not to affect the efficiency with which the AI weighing system operates.
  • the learning unit 2132 can perform learning based on the data stored in the memory 212, and create and train an AI model.
  • the learning unit 2132 can also use the data generated during the application process and other learning data (for example, data in the cloud environment) to retrain the recognition model.
  • the learning part 2132 can also train the learning model by using learning data to change the weights of various parameters in the AI inference model; and/or by preparing multiple AI inference models and then changing the AI inference model to be used according to the content of the calculation process.
  • the training of the AI inference model by the learning unit 2132 is preferably performed when the AI weighing system is idle.
  • the AI processing unit 213 of the image sensor 21 may only include the inference unit 2131 .
  • the function of the learning part 2132 is executed on the cloud environment when the AI weighing system is idle and when the network is available.
  • the AI processing unit 213 of the image sensor 21 can use the AI inference model stored in the memory 212 to only perform the AI recognition of the target, thereby further improving the recognition speed, and completely independent of network connection and cloud environment.
  • the learning part of the AI processing part according to the present invention can be located in any external computing device of the AI weighing system accessible in a wired or wireless manner, such as cloud environment (cloud), edge server, core network, etc.
  • a cloud server located in a cloud environment may include one or more computing devices with corresponding processors and memory capable of high-speed processing of large amounts of (and updated in real time) data.
  • the AI weighing system executes the training of the inference model ("the second inference model”) on the cloud environment when the system is idle and the network is available, and automatically deploys the inference model ("the second inference model") to the memory 212 of the image sensor 21, for example periodically, to update the inference model ("the first inference model”) in the memory 212.
  • the cloud environment can acquire, store and process large amounts of data, it is conducive to the establishment and retraining of AI inference models.
  • the AI inference model on the system can be updated when the network is available or the AI weighing system is idle, thereby improving the recognition accuracy without affecting the recognition speed when the system is working.
  • image data generated by image sensor 21 is also sent and stored to one or more computing devices in the cloud.
  • the computing device located in the cloud environment can create and/or retrain the AI inference model based on the image data generated by the image sensor 21, and can directly deploy the AI inference model to the memory 212 of the image sensor 21, thereby updating the AI inference model stored in the image sensor 21.
  • the AI processing unit 213 sends the recognition result of the target to the output unit 3 .
  • the recognition result may include N options (N ⁇ 1), and when N>1, the N options are arranged according to the ranking of possibility.
  • the output unit 3 displays three possible results of the target object 4 , and the first option is the best result recognized by the system. If the best result is not the correct result for the object 4 , the user can select the correct result on the interactive display screen of the output part 3 . This selection by the user will be used as historical data (learning data) for the feedback mechanism of the present invention.
  • the output unit 3 can also highlight the best result among multiple possible results by darkening the color, changing the font, enlarging the font size, and the like.
  • the image of the target used in the AI recognition process and the correct option are stored in the memory 212 of the image sensor 21 or sent to the cloud environment, and are fed back to the learning part 2132.
  • the learning part 2132 and/or the cloud processor associate the image of the target object and its correct options as learning data for retraining the AI inference model, so as to dynamically and continuously optimize the recognition accuracy of the model.
  • the image sensor 21 may be a stacked complementary metal oxide semiconductor (CMOS) image sensor chip.
  • the stacked CMOS image sensor chip includes a first substrate 301 and a second substrate 302 .
  • a pixel array part 3011 composed of a plurality of pixels is arranged on the first substrate 301, and the pixel array part 3011 converts optical signals into electrical signals through photoelectric conversion, and transmits them to the second substrate 302 (the connection between the first substrate and the second substrate is not shown).
  • a memory 3021 and a processing circuit 3022 are arranged on the second substrate 302 .
  • the processing circuit 3022 includes, for example, a DSP (Digital Signal Processor), which generates image data based on electrical signals transmitted from the first substrate 301, and stores the image data in the memory 3021 (for example, the memory 212 shown in FIG. 2 ).
  • the memory 3021 stores an AI model, for example, an inference model for AI recognition of a target.
  • the processing circuit 3022 executes the AI processing function of the inference part 2131 of the AI processing part 213 in FIG.
  • the processing circuit 3022 can also perform the function of the learning unit 2132 in FIG. 2 as described above.
  • the stacked CMOS image sensor chip may further include a third substrate.
  • the memory and processing circuitry may be located on the second substrate and/or the third substrate, respectively. That is, the pixel array unit, the memory, and the processing unit may be respectively located on different substrates.
  • the image sensor 21 of the present invention can itself perform AI recognition processing on the image of the target object, thereby enabling recognition of the target object in an offline state.
  • Fig. 4 illustrates the steps of creating or retraining an AI model according to the present invention.
  • the AI processing part specifically, the learning part 2132 located in the cloud environment or image sensor
  • acquires learning data the learning data includes the data set of the item image and the name and attribute of the item, etc., and the learning data also includes historical identification data of the AI weighing system
  • the learning part 2132 uses the learning data to create an AI model, or trains or retrains the created AI model
  • the created or updated AI model is output.
  • the image data used to create or train the AI model includes a variety of different image data sets, such as outline images of items, images of items with packaging, cross-sectional images of items, partial images of items, and the like.
  • the AI model of the present invention is created using more than two kinds of data sets from the above-mentioned data sets.
  • step S01 multiple sets of images of the overall outline of the item (data set 1), such as the profile of the item taken from different angles; multiple sets of images of the packaged item (data set 2), such as images of items in packaging of various colors (for example, various common plastic bags); multiple sets of images of different sections of the item (data set 3), such as images of items segmented into different shapes; and local enlarged images of the item (data set 4).
  • data set 1 For different items, you can choose to use different data sets for modeling. For example, for watermelon and other commodities that may be sold separately, two data sets (ie, data sets 1 and 3) can be collected for the overall image and the segmented cross-sectional image.
  • the product when a customer chooses to buy, for example, half a watermelon, the product can be quickly and accurately identified.
  • two data sets namely, data sets 1 and 4
  • the grapes can be quickly identified by confirming the local details.
  • the item image data used to create or train the AI model described above is generally an RGB image or a black and white image.
  • the image data used for creating or training the AI model may also include one or more of three-dimensional data, polarization data, and multi-wavelength data of the object.
  • the ToF sensor can be used to obtain the three-dimensional data of the item, that is, the ToF data (data set 5), and it is fused with the RGB image of the item (at least one of the data sets 1 to 4) to obtain a stereo image of the item.
  • Stereo images can fully present the surface features of objects, thereby optimizing the recognition accuracy of AI models.
  • the polarized image of the item obtained by a polarized light sensor (Dataset 6) can be used when modeling. By using polarized images, it is possible to avoid the problem that the captured image is unclear due to reflections on the surface of the object.
  • the item can also be photographed using a multi-wavelength sensor to obtain a multi-wavelength image of the item (data set 7).
  • multi-wavelength sensors are able to capture small differences in the surface color of items
  • the combined use of multi-wavelength images with RGB images of items is beneficial for modeling for identifying the presence of many species of the same item (e.g., the same fruit with different origins).
  • the AI model of the present invention can significantly improve the recognition accuracy compared with the conventional recognition model. Furthermore, the present invention further improves the recognition accuracy of the AI model by making the image data used for modeling include other image data (data sets 5-7) other than RGB images.
  • the photographing device according to the present invention zooms in and processes the part of the item based on optical zoom instead of digital zoom commonly used in the prior art.
  • the acquisition of the partially enlarged image of the target by the shooting device according to the present invention is also based on optical zoom. Therefore, the image quality is not degraded by enlarging the image, thereby further improving the recognition accuracy. This will be described in detail below with reference to FIG. 6 .
  • Fig. 5 shows the steps of the method for AI identification of a target by using the AI weighing identification system of the present invention.
  • step S501 after sensing that the target object 4 is placed on the storage table 1, the camera 2 takes pictures of the target object and acquires multiple images of the target object, including overall outline images, partial images, etc., as will be described below with reference to FIG. 6 .
  • the camera device 2 can also acquire image data other than the RGB image of the target through other sensors, such as ToF data, polarized light images and multi-wavelength images.
  • the AI processing unit 213 in the imaging device 2 performs AI processing on the target image to obtain high-quality image data for AI recognition.
  • the AI processing here includes intercepting the local detail image of the target object from the captured image, as described below with reference to FIG. 6 .
  • the AI processing of the image further includes fusing or combining the RGB image with other sensor data such as a ToF sensor to optimize the image data of the target object.
  • Fig. 6 shows an image output and/or processing mode of an image sensor according to the present invention.
  • the size of the image output by the image sensor 21 according to the present invention can be selected from a variety of pixel sizes. For example, 4056x3040 pixels (12M full sensor size), 1947x1459 pixels (covering the entire shelf) or 640x480 pixels (Video Graphics Array (VGA) size) shown in Figure 6. Therefore, according to the control of the control unit, the imaging unit 211 of the image sensor 21 can choose to capture a panoramic image (mode 1), an overall image of the target object (mode 2) and a partial image (mode 3). Mode 3 can optimize the recognition result when the target object is, for example, grapes.
  • the AI processing unit 213 can also cut out an image of the object area from the full sensor size image, and can cut out a VGA size image from the full sensor size image or the object area image. In this way, it is also possible to acquire local detailed images of objects.
  • the present invention can utilize an image sensor with an AI processing function to output a partially enlarged image of a target object.
  • the above-mentioned image enlargement process according to the present invention is based on optical zoom, so the quality of the enlarged image will not be affected at all, thereby greatly improving the recognition accuracy and efficiency compared with the prior art.
  • the data volume of the image captured by the camera device of the present invention is larger than that of the traditional camera device, since the image sensor of the present invention itself has AI processing and recognition functions, instead of sending image data to the cloud for processing and recognition as in the prior art, the recognition speed will not be affected, and the risk of loss or error during data transmission will be avoided. Therefore, by using high-quality images, the present invention greatly improves the recognition accuracy of the target object without affecting the recognition speed.
  • the AI processing unit 213 uses the AI inference model in the memory 212 to recognize the target object based on the image data processed in step S502 .
  • the recognition result may include N (N ⁇ 1) options.
  • the AI processing unit 213 may weight different options in combination with other obtained information, so as to optimize the ranking of multiple options.
  • Other information refers to information that may affect the user's (for example, a customer in a supermarket or a store)'s choice of goods, such as the identified user's anonymous characteristic data (such as gender and age, etc.), the address information of the system's location, weather and temperature information, etc.
  • step S504 the AI processing unit 213 outputs the recognition result to the output unit 3 .
  • the output unit 3 displays the recognition result to the user. If the first option is not the correct result, the user can select the correct result in other options, or manually input the correct result.
  • step S505 the AI processing unit 213 stores the image data and the correct recognition result in the memory 212, and/or uploads them to the memory of the computing device in the cloud environment. As mentioned earlier, this data can be used to retrain the AI inference model according to the feedback mechanism.
  • Fig. 7 shows a second embodiment of the AI weighing system according to the present invention.
  • the image sensor 21 of the imaging device 2 is capable of outputting RGB color images and/or black and white images.
  • the imaging device 2 further includes other sensors such as a ToF sensor 22 , a polarization sensor 23 and a multi-wavelength sensor 24 . These other sensors are all connected to the image sensor 21 .
  • the AI processing unit 213 can combine or fuse the data output by various types of sensors for AI recognition, so as to optimize the image of the target object and increase the recognition accuracy of the target object.
  • the ToF sensor 22 is a ToF (Time of Flight) distance image sensor that measures the distance to the target by detecting the flight time (time difference) of the light emitted by the light source being reflected by the target and reaching the sensor.
  • the AI processing unit 213 can fuse the RGB image obtained from the imaging unit 211 with the ToF image output from the ToF sensor 22 to obtain a stereoscopic image of the target. Recognition based on the stereo image can increase the recognition accuracy of the target.
  • the polarization sensor 23 is, for example, a sensor obtained by incorporating a polarization element, which is an independent component in a conventional polarization camera, into a CMOS image sensor.
  • the polarized light sensor 23 can capture objects that cannot be seen clearly due to light reflection. For example, plastic bags for packaging vegetables will show uneven reflection, and the unevenness of the surface of the object will be displayed delicately, thereby optimizing the captured image of the target object.
  • the AI processing unit 213 can choose to use the output of the polarization sensor to optimize the recognition accuracy of the target object.
  • the multi-wavelength sensor 24 can capture small differences in the color of the target object by utilizing different sensitivities to different wavelengths of light.
  • the AI processing unit 213 is particularly advantageous when identifying fruits such as citrus by utilizing the difference in sensitivity of the target object to different wavelengths output by the multi-wavelength sensor 24 .
  • the AI processing unit 213 can selectively combine or fuse one or more of the data output by the above sensors with the data output by the image sensor 21 .
  • the second embodiment further improves the recognition effect by using multiple types of sensor data individually or in combination.
  • Fig. 8 shows a third embodiment of the AI weighing system according to the present invention.
  • the difference between the third embodiment and the first embodiment is that two imaging devices are provided.
  • the AI weighing system according to this embodiment includes a camera 2 and a camera 5 communicating with the camera 2 .
  • the imaging device 2 is the same as that of the first embodiment.
  • the camera device 5 is arranged at a different position from the camera device 2 on the shelf, for example, in front of the checkout counter of a supermarket or the like. As shown in FIG. 8 , the imaging device 5 is used to capture an image of a target person (for example, a customer), and transmit the captured image to the imaging device 2 .
  • the AI processing unit 213 of the imaging device 2 can acquire anonymous personal information (characteristic data) of the customer, such as metadata representing gender and age, from the image of the customer.
  • the AI processing section 213 may only store and output the anonymous personal information without saving the image of the customer.
  • the AI processing unit 213 assists in identifying the target based on the preference information associated with the anonymous personal information, for example, by changing the weight of an option in the identification result.
  • the camera 5 itself may also include an image sensor having an AI processing function.
  • the image sensor processes the image of the customer to obtain characteristic data of the customer (for example, the customer's gender and age).
  • the camera 5 sends only the metadata obtained from the captured image, not the image itself, to the camera 2, thereby fully protecting personal privacy while increasing transmission and processing speed.
  • the camera device 5 may include an RGB image sensor and a ToF sensor for respectively obtaining the RGB image and the ToF image of the customer, and sending the two images to the AI processing unit 213 of the camera device 2 .
  • the AI processing unit 213 combines the RGB image and the ToF image of the customer to generate a stereoscopic image of the customer, and obtains characteristic data of the customer based on the stereoscopic image.
  • the camera 5 itself includes an image sensor with an AI processing function
  • the above-mentioned AI processing (including generating a stereoscopic image, and feature analysis based on the stereoscopic image) will be performed at the camera 5 side, and the camera 5 only sends the processing result to the camera 2.
  • the camera device 5 obtains the customer as a woman about 25 years old from the captured image of the customer, and sends the metadata (ie, "female, 25 years old") to the camera device 2 .
  • the imaging device 2 stores the metadata in the memory 212 and sends it to the inference unit 2131 .
  • the preference information of a 25-year-old female is stored in the memory 212, and the preference information will be regularly updated according to historical data.
  • the inference unit 2131 uses the AI model to obtain a preliminary recognition result of the object based on the image of the object captured by the imaging unit 211, which may be orange, tangerine, tomato or lemon.
  • the inference unit 2131 learns that women around the age of 25 prefer sweeter oranges, and uses this as auxiliary identification information. In the absence of other recognition parameters, the inference unit 2131 will finally use oranges as the first-ranked option in the recognition results according to the preference. After the customer's payment is completed, the above metadata and the final result of the target object (whether the user selected the first ranked orange) can be used to retrain the AI model.
  • the camera 5 can also take an image of the shopping basket or shopping cart held by the customer, and send the image data to the camera 2.
  • the camera device 2 can pre-recognize the image data obtained from the camera device 5 .
  • the image data captured by the imaging device 5 can be subjected to AI processing for pre-identification, and the metadata of the processing result can be directly sent to the imaging device 2 .
  • camera device 5 may send a list of objects identified from the image data to camera device 2 .
  • the pre-recognition can be used as auxiliary information for the final AI recognition performed by the imaging device 2 .
  • the metadata for example, gender and age range, etc.
  • the recognition accuracy of the AI model is further improved.
  • the AI processing unit 213 can also identify the customer's QR code (two-dimensional code) and obtain anonymous information such as the customer's purchase history for updating preference information and retraining the AI model.
  • QR code two-dimensional code
  • Fig. 9 shows a fourth embodiment of the AI weighing system according to the present invention.
  • the difference between the fourth embodiment and the third embodiment is that an imaging device 6 is also provided, that is, this embodiment includes three imaging devices 2 , 5 and 6 .
  • the imaging devices 2 and 5 are the same as those of the second embodiment.
  • the imaging device 6 is provided at a different position from the imaging devices 2 and 5 .
  • a plurality of camera devices 6 are respectively installed at a plurality of different positions in a place such as a supermarket.
  • FIG. 9 shows an imaging device 6 installed in a refrigerator in a supermarket.
  • the camera device 6 can be set at each commodity display place of the supermarket.
  • the imaging device 6 photographs customers within the imaging range, and sends the captured image to the imaging device 2 .
  • the camera device 2 can perform simultaneous localization and mapping (SLAM) processing on the image of the customer.
  • SLAM simultaneous localization and mapping
  • the AI processing unit 213 analyzes the feature points in the image, determines whether a specific feature point moves by a certain vector relative to another image, and generates a SLAM map by combining feature point data of multiple images taken successively. In this way, the AI processing unit 213 can determine the product information that the customer takes out or puts down at the container where the camera device 6 is installed, in combination with the product information stored in the system memory at the container where the camera device 6 is installed.
  • the camera device 6 can also take pictures of customers' shopping carts or shopping baskets within the shooting range, and send the captured images to the camera device 2 .
  • the AI processing unit 213 can perform SLAM processing on the image of the shopping cart or shopping basket, so as to determine the product information put into or taken away from the shopping cart or shopping basket.
  • the AI processing unit 213 can also compare the images of shopping carts or shopping baskets captured by multiple camera devices 6 arranged at different locations, and determine the product information put into or taken away from the shopping cart or shopping basket through image differences.
  • the camera 6 can send the overall image of the customer and his shopping cart or shopping basket to the camera 2, and the AI processing unit 213 can determine the product information that the customer puts into the shopping cart or shopping basket or takes away from the shopping cart or shopping basket based on the overall image.
  • the camera 6 itself may also include an image sensor with an AI processing function, and the AI image sensor performs the aforementioned AI processing on the camera 6 and sends metadata of the processing result to the camera 2 .
  • the AI image sensor performs the aforementioned AI processing on the camera 6 and sends metadata of the processing result to the camera 2 .
  • a customer takes out/puts down a product from a container at a certain time, a customer's shopping cart or basket is put in or taken out a certain product at a certain time.
  • the imaging device 2 can process the image of the customer captured by the imaging device 6 to obtain the customer's anonymous information (for example, gender, age, etc.).
  • the imaging device 6 itself has an AI image sensor, only the metadata of the above information can be sent to the AI processing unit 213 of the imaging device 2 .
  • the AI processing unit 213 can associate the customer's metadata with the commodity data selected by him and store it in the memory 212 for retraining the AI model.
  • the metadata for example, gender and age range, etc.
  • the recognition accuracy of the AI model is further improved.
  • Fig. 10 shows a fifth embodiment of the AI weighing system according to the present invention.
  • multiple AI weighing and payment systems may be installed in the same store, and multiple systems are connected to each other in a wireless or wired manner.
  • multiple AI weighing systems communicate with each other to update the store's database (including product information, customer's anonymous personal information and purchase history data, etc.).
  • the AI model of each AI image sensor is retrained to continuously improve the recognition accuracy of the AI model.
  • the training and retraining of the AI model can also be performed on the computing device in the cloud environment, and the updated AI model can be deployed to each AI weighing system.
  • database information of different stores can also be uploaded to computing devices in the cloud environment.
  • the AI processing units of the AI weighing systems in different stores can also communicate in real time.
  • each store can obtain other information, such as real-time temperature, area address, and local climate, etc., from the cloud environment, for example. Utilizing these information can also assist in improving the recognition accuracy of the AI processing unit for the target object (for example, a product in a store).
  • the application scenario of the present invention is described above by taking the weighing of goods in a supermarket as an example.
  • the present invention is obviously not limited to this.
  • the AI recognition system of the present invention can also be applied to any scene where commodities/items need to be recognized, and is not limited to weighing scenes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Mathematical Physics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Image Analysis (AREA)
  • General Factory Administration (AREA)

Abstract

The present invention relates to an AI weighing system, and a method for improving the precision of an AI model by using various types of data sets. The AI weighing system may comprise: an object placement table, which is used for the placement of a target object, wherein the object placement table can weigh the target object; a first photographing apparatus, which is used for photographing and recognizing the target object, which is placed on the object placement table, wherein the first photographing apparatus comprises an image sensor, and the image sensor can execute, in an offline state, AI processing for recognizing the target object; and an output portion, which is used for outputting a recognition result and a weighing result of the target object. The AI weighing system according to the present invention significantly improves the recognition speed and precision.

Description

AI称重系统和利用多种数据集增加AI模型的精度的方法AI weighing system and method to increase the accuracy of AI models using multiple data sets
相关申请的引用References to related applications
本申请要求于2022年01月20日向中华人民共和国国家知识产权局提交的第202210065055.5号中国专利申请的权益,在此将其全部内容以援引的方式整体并入本文中。This application claims the rights and interests of the Chinese patent application No. 202210065055.5 submitted to the State Intellectual Property Office of the People's Republic of China on January 20, 2022, the entire contents of which are hereby incorporated herein by reference.
技术领域technical field
本发明涉及AI(Antificial Intelligence)称重系统,更具体地,涉及利用具有AI处理功能的图像传感器的AI称重(AI scale)系统。The present invention relates to an AI (Antificial Intelligence) weighing system, more specifically, relates to an AI weighing (AI scale) system utilizing an image sensor with an AI processing function.
背景技术Background technique
在例如超市等的销售场所,已经出现了顾客能够自助结算和支付的工具。例如,在超市中,顾客能够利用商品上的条形码自行扫描和结算。然而,条形码仍然需要人工粘附在商品上。尤其是,对于例如生鲜类商品,还需要人工识别和称重,再打印出条形码。由于商品种类繁多,人工识别商品耗时、费力且容易出现错误。In sales places such as supermarkets, there have been tools for customers to self-checkout and pay. For example, in a supermarket, customers can use the barcode on the product to scan and pay by themselves. However, barcodes still need to be manually attached to merchandise. Especially, for such as fresh commodities, manual identification and weighing are also required, and then barcodes are printed out. Due to the wide variety of products, manual product identification is time-consuming, laborious and error-prone.
目前,也出现了基于计算机视觉技术的人工智能(AI)称重装置。这样的AI称重装置通过建立智能识别模型并利用模型分析商品图像来自动地识别商品。这样的AI称重装置避免了人工称重和打印条形码的步骤,且避免了需要人工查找或记忆各种商品的价格,从而极大改善了诸如超市这样的场所的工作效率且节省了人工费用。Currently, there are also artificial intelligence (AI) weighing devices based on computer vision technology. Such an AI weighing device automatically recognizes a product by establishing an intelligent recognition model and analyzing a product image using the model. Such an AI weighing device avoids the steps of manual weighing and printing barcodes, and avoids the need to manually search or memorize the prices of various commodities, thereby greatly improving the work efficiency of places such as supermarkets and saving labor costs.
发明内容Contents of the invention
然而,在现有的AI称重装置中,对商品的识别是通过将照相机拍摄的图像发送到诸如云端的外部计算设备上进行的,并且AI模型的学习、建立、训练和再训练同样是在云端进行。因而,这存在至少以下三种问题。首先,数据传输取决于数据流量(data traffic)可能是不稳定的且存在延迟甚至错漏,从而显著影响识别速度和精度。其次,云端AI识别依赖于网络的稳定性,且在离线(off line)时无法实施。因此,云端AI识别显著依赖于云环境。第三,发往云端的图片可能包括隐私信息,例如用户的个人信息,从而存在隐私泄露的风险。However, in the existing AI weighing device, the identification of goods is carried out by sending the images captured by the camera to an external computing device such as the cloud, and the learning, establishment, training and retraining of the AI model are also carried out on the cloud. Thus, there are at least the following three problems. First of all, data transmission depends on data traffic, which may be unstable and have delays or even errors, which will significantly affect the recognition speed and accuracy. Secondly, cloud AI recognition depends on the stability of the network and cannot be implemented offline. Therefore, cloud AI recognition significantly depends on the cloud environment. Third, the pictures sent to the cloud may contain private information, such as the user's personal information, which poses a risk of privacy leakage.
同时,现有AI称重装置的识别模型的识别精度有待改善。尤其是,在诸如超市这样的场所中,即使是同一种类的商品,其包装和出售形式等也多种多样。例如,对于生鲜类商品,其存在盒装、塑料袋等多种包装形式,还可能自带包装;同时,其可能完整出售,也可能分割出售(例如,切片、分块等等)。这些都给精准识别商品带来了难度。At the same time, the recognition accuracy of the recognition model of the existing AI weighing device needs to be improved. In particular, in a place such as a supermarket, even the same type of product has various packages, sales forms, and the like. For example, for fresh goods, there are various packaging forms such as boxed and plastic bags, and may also be packaged by itself; at the same time, they may be sold whole or divided (for example, sliced, divided into pieces, etc.). All of these make it difficult to accurately identify products.
为解决上述问题,本发明通过利用具有AI处理功能的图像传感器,能够在离线状态执行AI识别,从而显著改善了智能称重装置的识别精度、识别速度,并且能够离线使用。In order to solve the above problems, the present invention can perform AI recognition offline by using an image sensor with AI processing function, thereby significantly improving the recognition accuracy and recognition speed of the intelligent weighing device, and can be used offline.
本发明还通过利用包含多种数据集的图像数据显著改善了AI模型的识别精度。The present invention also significantly improves the recognition accuracy of the AI model by utilizing image data containing various data sets.
根据本发明的AI称重系统可包括:置物台,用于放置目标物,所述置物台能够对所述目标物进行称重;第一摄像装置,用于对放置在所述置物台上的所述目标物进行拍摄和识别,所述第一摄像装置包括图像传感器,所述图像传感器能够在离线状态下执行用以识别所述目标物的AI处理;以及输出部,用于输出所述目标物的识别结果以及称重结果。The AI weighing system according to the present invention may include: a storage platform for placing a target object, the storage platform can weigh the target object; a first camera device for photographing and identifying the target object placed on the storage platform, the first camera device includes an image sensor, and the image sensor can perform AI processing for identifying the target object in an offline state; and an output unit is used for outputting the recognition result and the weighing result of the target object.
根据本发明的AI称重系统的摄像装置的图像传感器可为CMOS图像传感器芯片,所述CMOS图像传感器芯片包括第一基板和第二基板,所述第一基板具有将光信号转化为电信号的多个像素,且所述第二基板具有存储器和处理电路,所述存储器存储有AI模型,所述处理电路具有通过使用所述AI模型基于所述电信号来执行所述AI处理的功能。The image sensor of the camera device of the AI weighing system according to the present invention may be a CMOS image sensor chip, the CMOS image sensor chip includes a first substrate and a second substrate, the first substrate has a plurality of pixels for converting optical signals into electrical signals, and the second substrate has a memory and a processing circuit, the memory stores an AI model, and the processing circuit has a function of performing the AI processing based on the electrical signal by using the AI model.
图像传感器上存储的AI模型包括第一推论模型。The AI models stored on the image sensor include a first inference model.
根据本发明的一方面,堆叠式CMOS图像传感器芯片的所述处理电路生成图像数据,并且所述处理电路可包括:学习部,所述学习部基于所述图像数据再训练所述AI模型;以及推论部,所述推论部利用所述AI模型对所述目标物进行识别。According to an aspect of the present invention, the processing circuit of the stacked CMOS image sensor chip generates image data, and the processing circuit may include: a learning part that retrains the AI model based on the image data; and an inference part that uses the AI model to recognize the target object.
根据本发明的另一方面,AI称重系统还可包括位于云环境中的一个或多个计算设备,所述一个或多个计算设备具有相应的处理器和存储器。其中,第一摄像装置的堆叠式CMOS图像传感器芯片的所述处理电路生成图像数据,并且所述图像数据被发送至所述一个或多个计算设备。所述一个或多个计算设备基于所述堆叠式CMOS图像传感器芯片生成的所述图像数据创建第二推论模型,并且可直接将所述第二推论模型部署到所述堆叠式CMOS图像传感器芯片的所述存储器中,使得所述第一推论模型被更新。According to another aspect of the present invention, the AI weighing system may further include one or more computing devices located in a cloud environment, and the one or more computing devices have corresponding processors and memories. Wherein the processing circuit of the stacked CMOS image sensor chip of the first camera generates image data, and the image data is sent to the one or more computing devices. The one or more computing devices create a second inference model based on the image data generated by the stacked CMOS image sensor chip, and may directly deploy the second inference model into the memory of the stacked CMOS image sensor chip such that the first inference model is updated.
根据本发明的另一方面,所述堆叠式CMOS图像传感器芯片能够选择所述图像数据的尺寸,所述尺寸包括全传感器尺寸和基于所述处理电路的所述AI处理的视频图形阵列(VGA)尺寸。According to another aspect of the present invention, the stacked CMOS image sensor chip is capable of selecting a size of the image data including a full sensor size and a Video Graphics Array (VGA) size based on the AI processing of the processing circuit.
根据本发明的另一方面,当所述堆叠CMOS图像传感器输出具有全传感器尺寸的所述图像数据时,所述AI处理包括从具有所述全传感器尺寸的所述图像数据中截取所述目标物的整体图像或局部图像,所述局部图像包括VGA尺寸图像。According to another aspect of the present invention, when the stacked CMOS image sensor outputs the image data having a full sensor size, the AI processing includes intercepting an overall image or a partial image of the target object from the image data having the full sensor size, and the partial image includes a VGA-sized image.
根据本发明的另一方面,AI称重系统的第一摄像装置拍摄的图像数据可包括:所述目标物的轮廓图像;和/或所述目标物的截面图像;和/或所述目标物的包装图像。According to another aspect of the present invention, the image data captured by the first camera device of the AI weighing system may include: a profile image of the target object; and/or a cross-sectional image of the target object; and/or a packaging image of the target object.
根据本发明的另一方面,AI称重系统的第一摄像装置拍摄的图像数据还可包括所述目标物的局部放大图像。According to another aspect of the present invention, the image data captured by the first camera device of the AI weighing system may also include a partially enlarged image of the target object.
根据本发明的另一方面,AI称重系统的第一摄像装置还可包括ToF传感器,并且所述AI处理包括将所述图像传感器输出的RBG数据和所述ToF传感器输出的ToF数据进行组合。According to another aspect of the present invention, the first camera device of the AI weighing system may further include a ToF sensor, and the AI processing includes combining RBG data output by the image sensor and ToF data output by the ToF sensor.
根据本发明的另一方面,AI称重系统的第一摄像装置还可包括多波长传感器,并且所述AI处理能够利用所述多波长传感器的输出优化对所述目标物的识别。According to another aspect of the present invention, the first camera device of the AI weighing system may further include a multi-wavelength sensor, and the AI processing can utilize the output of the multi-wavelength sensor to optimize the identification of the target object.
根据本发明的另一方面,AI称重系统的第一摄像装置还可包括偏光传感器,并且所述AI处理能够利用所述偏光传感器的输出优化对所述目标物的识别。According to another aspect of the present invention, the first camera device of the AI weighing system may further include a polarized light sensor, and the AI processing can utilize the output of the polarized light sensor to optimize the identification of the target object.
根据本发明的另一方面,AI称重系统还可包括:第二摄像装置,用于获取目标人物的图像并发送给所述第一摄像装置,并且其中,所述AI处理包括从所述第二摄像装置获取的所述目标人物的图像获取特征数据并且利用所述特征数据辅助识别所述目标物。其中,所述特征数据包括所述目标人物的性别和年龄等匿名的特征数据。According to another aspect of the present invention, the AI weighing system may further include: a second camera device for capturing an image of the target person and sending it to the first camera device, and wherein the AI processing includes acquiring feature data from the image of the target person captured by the second camera device and using the feature data to assist in identifying the target object. Wherein, the feature data includes anonymous feature data such as gender and age of the target person.
根据本发明的另一方面,AI称重系统还可包括:第三摄像装置,用于获取目标人物的图像并发送给所述第一摄像装置,并且其中,所述AI处理包括对所述第三摄像装置获取的所述目标人物的图像进行SLAM处理,并输出处理结果的元数据以辅助识别所述目标物。其中,所述第三摄像装置还获取所述目标人物的购物车或购物筐的图像并发送给所述第一摄像装置,并且所述AI处理包括对所述购物车或购物筐的图像进行SLAM处理,并输出处理结果的元数据以辅助识别所述目标物。According to another aspect of the present invention, the AI weighing system may further include: a third camera device, configured to acquire an image of the target person and send it to the first camera device, and wherein the AI processing includes performing SLAM processing on the image of the target person captured by the third camera device, and outputting metadata of the processing result to assist in identifying the target object. Wherein, the third camera device also acquires an image of the shopping cart or shopping basket of the target person and sends it to the first camera device, and the AI processing includes performing SLAM processing on the image of the shopping cart or shopping basket, and outputting metadata of the processing result to assist in identifying the target object.
优选地,本发明的第三摄像装置可包括定位在所述目标人物的移动范围内的不同位置处的多个图像传感器。Preferably, the third camera device of the present invention may include a plurality of image sensors positioned at different positions within the moving range of the target person.
根据本发明的另一方面,AI称重系统的AI处理还可包括获取包括环境温度、区域地址和/或气候条件的其它信息,并能够基于所述其它信息辅助识别所述目标物。According to another aspect of the present invention, the AI processing of the AI weighing system may also include obtaining other information including ambient temperature, area address and/or weather conditions, and assist in identifying the target object based on the other information.
根据本发明的利用多种数据集来增加AI模型的识别精度的方法,包括:获取物品的图像数据;利用学习数据创建AI模型,所述学习数据包括所述图像数据和所述物品的名称以及属性;应用所述AI模型识别目标物,其中,所述学习数据包括以下三种数据集中的至少两种:所述物品的轮廓图像;所述物品的截面图像;所述物品的包装图像。The method for increasing the recognition accuracy of an AI model by using multiple data sets according to the present invention includes: acquiring image data of an item; creating an AI model using learning data, the learning data including the image data and the name and attribute of the item; applying the AI model to identify the target object, wherein the learning data includes at least two of the following three data sets: an outline image of the item; a cross-sectional image of the item; and a packaging image of the item.
优选地,用于创建AI模型的学习数据还可包括物品的局部放大图像的数据集。Preferably, the learning data used to create the AI model may also include a data set of partially enlarged images of the item.
优选地,所述物品的所述图像数据包括RGB图像数据以及以下数据中的至少一者:ToF数据、多波长数据、偏光数据。Preferably, the image data of the item includes RGB image data and at least one of the following data: ToF data, multi-wavelength data, polarization data.
通过本发明的上述一个或多个方面,AI称重系统的识别速度和精度得到显著改善,并且尤其是,即使在离线状态下也能够精确且快速地识别目标物。并且,根据本发明的创建AI模型的方法通过增加不同的数据集显著改善了识别模型的精度。Through the above one or more aspects of the present invention, the recognition speed and precision of the AI weighing system are significantly improved, and especially, the target object can be recognized accurately and quickly even in an offline state. Also, the method for creating an AI model according to the present invention significantly improves the accuracy of the recognition model by adding different data sets.
附图说明Description of drawings
在以下说明中,将参照附图更全面地公开各种实施例的这些及其他更详细、具体的特征,其中:These and other more detailed, specific features of various embodiments will be more fully disclosed in the following description with reference to the accompanying drawings, in which:
图1是示出根据本发明的AI称重系统的第一实施方式的示意图。FIG. 1 is a schematic diagram showing a first embodiment of an AI weighing system according to the present invention.
图2是示出根据本发明的图像传感器的实施例的示意图。FIG. 2 is a schematic diagram showing an embodiment of an image sensor according to the present invention.
图3是示出根据本发明的堆叠式CMOS图像传感器芯片的示意图。FIG. 3 is a schematic diagram showing a stacked CMOS image sensor chip according to the present invention.
图4是示出根据本发明的用于创建、训练或再训练AI模型的步骤的流程图。FIG. 4 is a flowchart illustrating steps for creating, training or retraining an AI model according to the present invention.
图5是示出根据本发明的AI识别方法的流程图。FIG. 5 is a flowchart illustrating an AI recognition method according to the present invention.
图6是示出根据本发明的图像传感器的图像输出和/或处理模式的示意图。FIG. 6 is a schematic diagram illustrating an image output and/or processing mode of an image sensor according to the present invention.
图7是示出根据本发明的AI称重系统的第二实施方式的示意图。Fig. 7 is a schematic diagram showing a second embodiment of the AI weighing system according to the present invention.
图8是示出根据本发明的AI称重系统的第三实施方式的示意图。Fig. 8 is a schematic diagram showing a third embodiment of the AI weighing system according to the present invention.
图9是示出根据本发明的AI称重系统的第四实施方式的示意图。Fig. 9 is a schematic diagram showing a fourth embodiment of the AI weighing system according to the present invention.
图10是示出根据本发明的AI称重系统的第五实施方式的示意图10 is a schematic diagram showing a fifth embodiment of the AI weighing system according to the present invention
具体实施方式Detailed ways
在以下说明中,列出了许多细节,但是对于本领域技术人员来说显而易见的是,这些具体细节仅仅是示例性的,并不意图限制本申请的范围。In the following description, numerous details are set forth, but it will be apparent to those skilled in the art that these specific details are illustrative only and are not intended to limit the scope of the application.
<第一实施方式><First Embodiment>
如图1所示,根据本发明的第一实施方式的AI称重系统包括:置物台1、摄像装置2和输出部3,其中目标物4放置在置物台1上。置物台1内置称重装置,能够对目标物4进行称重。输出部3可为如图1所示的显示器,例如液晶显示器,以向用户显示目标物4的识别结果以及称重结果等。可选地,输出部3也可以以音频方式输出识别结果和称重结果。并且输出部3可为触摸显示器,以与用户进行交互。在本文中,以例如超市中售卖的商品(尤其是水果、蔬菜等生鲜产品)作为本发明的目标物进行了说明,但容易理解的是,本发明并不局限于这些特定商品。例如,本发明的目标物也可以是仓库中存储的货物。As shown in FIG. 1 , the AI weighing system according to the first embodiment of the present invention includes: a storage platform 1 , a camera device 2 and an output unit 3 , wherein an object 4 is placed on the storage platform 1 . The storage table 1 has a built-in weighing device, which can weigh the target object 4 . The output unit 3 can be a display as shown in FIG. 1 , such as a liquid crystal display, to display the recognition result and weighing result of the target object 4 to the user. Optionally, the output unit 3 may also output the recognition result and the weighing result in audio form. And the output unit 3 can be a touch display to interact with the user. In this article, for example, commodities sold in supermarkets (especially fresh products such as fruits and vegetables) are described as targets of the present invention, but it is easy to understand that the present invention is not limited to these specific commodities. For example, the object of the present invention may also be goods stored in a warehouse.
摄像装置2例如设置在置物台1上方,以对目标物进行拍摄。摄像装置2包括图像传感器21,例如CMOS图像传感器。根据本发明的图像传感器21不仅能够存储和输出拍摄的图像,并且能够对图像数据执行包括AI处理在内的各种处理。AI处理包括从图像数据中获得各种信息(诸如目标的特征数据等的元数据),以及对图像中的目标物进行识别等。如图2所示,图像传感器21也可与云环境(云端)进行通信。然而,由于图像传感器21本身可执行AI处理,因此,与现有的称重识别系统不同,本发明的AI称重系统并不依赖于网络连接和云端服务器,即使在离线状态,也能够快速且精确识别出目标物。The imaging device 2 is, for example, arranged above the object table 1 to photograph the target object. The imaging device 2 includes an image sensor 21 such as a CMOS image sensor. The image sensor 21 according to the present invention is capable of not only storing and outputting captured images, but also performing various processing including AI processing on image data. AI processing includes obtaining various information (metadata such as feature data of objects) from image data, and recognizing objects in images. As shown in FIG. 2 , the image sensor 21 can also communicate with a cloud environment (cloud). However, since the image sensor 21 itself can perform AI processing, unlike existing weighing recognition systems, the AI weighing system of the present invention does not depend on network connections and cloud servers, and can quickly and accurately identify objects even in an offline state.
[具有AI处理功能的图像传感器的结构][Structure of image sensor with AI processing function]
图2示出根据本发明的图像传感器21的功能框图。如图2所示,图像传感器21包括成像部211、存储器212和AI处理部213。图像传感器21还包括对成像部进行控制的控制部(未图示)。成像部211对目标物进行拍摄,并将成像数据发送到AI处理部213,并可将成像数据存储在存储器212中。FIG. 2 shows a functional block diagram of an image sensor 21 according to the invention. As shown in FIG. 2 , the image sensor 21 includes an imaging section 211 , a memory 212 and an AI processing section 213 . The image sensor 21 further includes a control unit (not shown) that controls the imaging unit. The imaging unit 211 photographs the target object, sends the imaging data to the AI processing unit 213 , and stores the imaging data in the memory 212 .
存储器212中存储有AI模型,例如对目标物进行识别的推论模型(“第一推论模型”)。AI模型例如为用于计算机视觉的神经网络计算模型,该模型通过执行例如存储在存储器212和/或云环境中的存储器中的程序而利用深度神经网络(DNN)的学习来创建。此外,AI模型可以是利用多层神经网络的学习模型。应当理解的是,这里的AI模型可以根据用途和需求选择任何合适的已知AI模型和算法。The memory 212 stores an AI model, for example, an inference model for recognizing a target ("first inference model"). The AI model is, for example, a neural network computing model for computer vision created using learning of a deep neural network (DNN) by executing a program, for example, stored in memory 212 and/or in a cloud environment. Also, the AI model may be a learning model utilizing a multi-layer neural network. It should be understood that for the AI model here, any suitable known AI model and algorithm may be selected according to usage and requirements.
AI处理部213例如为图形处理单元(Graphic Processing Unit,GPU),从而可利用存储器212中存储的AI推论模型对目标物的图像数据进行处理,并且将处理结果发送到输出部3。The AI processing unit 213 is, for example, a graphics processing unit (Graphic Processing Unit, GPU), so that the image data of the target object can be processed using the AI inference model stored in the memory 212, and the processing result is sent to the output unit 3.
优选地,AI处理部213可包括推论部2131和学习部2132。推论部2131利用存储器212中存储的AI推论模型对成像部211拍摄和发送的目标物图像进行识别,并将识别结果发送给输出部3。学习部2132基于该目标物图像以及确认后的识别结果对AI推论模型进行再训练。并且,学习部2132也可利用云环境中的数据对AI推论模型进行再训练。再训练可以在AI称重系统空闲时和/或可选择在网络可用时进行,从而不影响AI称重系统进行操作时的效率。Preferably, the AI processing unit 213 may include an inference unit 2131 and a learning unit 2132 . The inference unit 2131 uses the AI inference model stored in the memory 212 to recognize the target image captured and sent by the imaging unit 211 , and sends the recognition result to the output unit 3 . The learning unit 2132 retrains the AI inference model based on the object image and the confirmed recognition result. In addition, the learning unit 2132 can also use data in the cloud environment to retrain the AI inference model. Retraining can be done when the AI weighing system is idle and/or optionally when the network is available, so as not to affect the efficiency with which the AI weighing system operates.
可选地,学习部2132可以基于存储器212中存储的数据进行学习,并且创建和训练AI模型。学习部2132也可利用应用过程中产生的数据以及其他学习数据(例如,云环境中的数据)对识别模型进行再训练。优选地,学习部2132还可以通过使用学习数据改变AI推论模型内的各种参数的权重来训练学习模型;和/或通过准备多个AI推论模型然后根据计算处理的内容改变待使用的AI推论模型。另外,如上所述的,学习部2132对AI推论模型的训练优选在AI称重系统空闲时进行。Optionally, the learning unit 2132 can perform learning based on the data stored in the memory 212, and create and train an AI model. The learning unit 2132 can also use the data generated during the application process and other learning data (for example, data in the cloud environment) to retrain the recognition model. Preferably, the learning part 2132 can also train the learning model by using learning data to change the weights of various parameters in the AI inference model; and/or by preparing multiple AI inference models and then changing the AI inference model to be used according to the content of the calculation process. In addition, as mentioned above, the training of the AI inference model by the learning unit 2132 is preferably performed when the AI weighing system is idle.
另外可选地,根据本发明的图像传感器21的AI处理部213可仅包括推论部2131。学习部2132的功能则在AI称重系统空闲时并且在网络可用时在云环境上执行。Alternatively, the AI processing unit 213 of the image sensor 21 according to the present invention may only include the inference unit 2131 . The function of the learning part 2132 is executed on the cloud environment when the AI weighing system is idle and when the network is available.
也就是说,图像传感器21的AI处理部213可利用存储器212中存储的AI推论模型仅执行目标物的AI识别,由此进一步提高了识别速度,且完全不依赖于网络连接和云环境。在这个方面,根据本发明的AI处理部的学习部可位于AI称重系统的以有线或无线方式可接入的任何外部计算设备中,例如云环境(云端)、边缘服务器、核心网络等等。位于云环境中的云服务器可包括一个或多个计算设备,其具有相应的处理器和存储器,可对大量(且实时更新的)数据进行高速处理。同时,AI称重系统在系统空闲时并且在网络可用时在云环境上执行推论模型(“第二推论模型”)的训练,并且例如定期自动地将该推论模型(“第二推论模型”)部署到图像传感器21的存储器212中,以更新存储器212中的推论模型(“第一推论模型”)。由于云环境能够获取、存储和处理大量数据,有利于AI推论模型的建立和再训练。同时,可在网络可用或AI称重系统空闲时更新系统上的AI推论模型,从而在不影响系统工作时的识别速度的情况下,改善识别精度。That is to say, the AI processing unit 213 of the image sensor 21 can use the AI inference model stored in the memory 212 to only perform the AI recognition of the target, thereby further improving the recognition speed, and completely independent of network connection and cloud environment. In this respect, the learning part of the AI processing part according to the present invention can be located in any external computing device of the AI weighing system accessible in a wired or wireless manner, such as cloud environment (cloud), edge server, core network, etc. A cloud server located in a cloud environment may include one or more computing devices with corresponding processors and memory capable of high-speed processing of large amounts of (and updated in real time) data. At the same time, the AI weighing system executes the training of the inference model ("the second inference model") on the cloud environment when the system is idle and the network is available, and automatically deploys the inference model ("the second inference model") to the memory 212 of the image sensor 21, for example periodically, to update the inference model ("the first inference model") in the memory 212. Since the cloud environment can acquire, store and process large amounts of data, it is conducive to the establishment and retraining of AI inference models. At the same time, the AI inference model on the system can be updated when the network is available or the AI weighing system is idle, thereby improving the recognition accuracy without affecting the recognition speed when the system is working.
另外,在这个方面,图像传感器21生成的图像数据也被发送且存储至云端的一个或多个计算设备中。位于云环境中的计算设备可基于图像传感器21生成的图像数据创建和/或再训练AI推论模型,并且可直接将该AI推论模型部署到图像传感器21的存储器212中,从而更新图像传感器21存储的AI推论模型。Additionally, in this aspect, image data generated by image sensor 21 is also sent and stored to one or more computing devices in the cloud. The computing device located in the cloud environment can create and/or retrain the AI inference model based on the image data generated by the image sensor 21, and can directly deploy the AI inference model to the memory 212 of the image sensor 21, thereby updating the AI inference model stored in the image sensor 21.
在完成AI识别后,AI处理部213将目标物的识别结果发送到输出部3。识别结果可能包括N个选项(N≥1),在N>1时,N个选项按可能性的等级(ranking)排列。如图1所示,输出部3显示目标物4的3种可能结果,第1选项为系统识别出的最佳结果。如果该最佳结果并非目标物4的正确结果,用户可在输出部3的交互显示屏上选择正确结果。用户的该选择将作为历史数据(学习数据)用于本发明的反馈机制。除了排序之外,输出部3在显示最佳结果时,还可以通过加深颜色、改变字体、放大字号等方式对多种可能结果中的最佳结果进行突出显示。After the AI recognition is completed, the AI processing unit 213 sends the recognition result of the target to the output unit 3 . The recognition result may include N options (N≥1), and when N>1, the N options are arranged according to the ranking of possibility. As shown in FIG. 1 , the output unit 3 displays three possible results of the target object 4 , and the first option is the best result recognized by the system. If the best result is not the correct result for the object 4 , the user can select the correct result on the interactive display screen of the output part 3 . This selection by the user will be used as historical data (learning data) for the feedback mechanism of the present invention. In addition to sorting, when displaying the best result, the output unit 3 can also highlight the best result among multiple possible results by darkening the color, changing the font, enlarging the font size, and the like.
具体地,根据本发明的反馈机制,当用户选择的最终选项(即,用户识别的目标物的结果)并非N个选项中的第1个选项时,AI识别过程中所使用的目标物图像以及该正确选项被存储到图像传感器21的存储器212或发送到云环境中,并且被反馈给学习部2132。学习部2132和/或云端处理器将该目标物图像及其正确选项关联起来作为学习数据,以用于对AI推论模型进行再训练,从而动态地、不断地优化模型的识别精度。Specifically, according to the feedback mechanism of the present invention, when the final option selected by the user (that is, the result of the target identified by the user) is not the first option among the N options, the image of the target used in the AI recognition process and the correct option are stored in the memory 212 of the image sensor 21 or sent to the cloud environment, and are fed back to the learning part 2132. The learning part 2132 and/or the cloud processor associate the image of the target object and its correct options as learning data for retraining the AI inference model, so as to dynamically and continuously optimize the recognition accuracy of the model.
[具有AI处理功能的图像传感器的构造][Construction of image sensor with AI processing function]
下面将根据图3来描述本发明的图像传感器的具体构造。如图3所示,图像传感器21可以是堆叠式互补金属氧化物半导体(CMOS)图像传感器芯片。堆叠式CMOS图像传感器芯片包括第一基板301和第二基板302。第一基板301上布置有由多个像素组成的像素阵列部3011,像素阵列部3011通过光电转换将光信号转换为电信号,并传输到第二基板302(第一基板和第二基板之间的连接未图示)。第二基板302上布置有存储器3021和处理电路3022。处理电路3022例如包括DSP(数字信号处理器,Digital Signal Processor),其基于从第一基板301传输的电信号生成图像数据,并将该图像数据存储在存储器3021(例如,图2所示的存储器212)中。存储器3021上存储有AI模型,例如对目标物进行AI识别的推论模型。处理电路3022执行如图2中的AI处理部213的推论部2131的AI处理功能,即,利用存储器212上存储的AI模型基于来自像素阵列部的电信号进行AI识别处理。处理电路3022还可以执行如上所述的图2中的学习部2132的功能。The specific configuration of the image sensor of the present invention will be described below with reference to FIG. 3 . As shown in FIG. 3 , the image sensor 21 may be a stacked complementary metal oxide semiconductor (CMOS) image sensor chip. The stacked CMOS image sensor chip includes a first substrate 301 and a second substrate 302 . A pixel array part 3011 composed of a plurality of pixels is arranged on the first substrate 301, and the pixel array part 3011 converts optical signals into electrical signals through photoelectric conversion, and transmits them to the second substrate 302 (the connection between the first substrate and the second substrate is not shown). A memory 3021 and a processing circuit 3022 are arranged on the second substrate 302 . The processing circuit 3022 includes, for example, a DSP (Digital Signal Processor), which generates image data based on electrical signals transmitted from the first substrate 301, and stores the image data in the memory 3021 (for example, the memory 212 shown in FIG. 2 ). The memory 3021 stores an AI model, for example, an inference model for AI recognition of a target. The processing circuit 3022 executes the AI processing function of the inference part 2131 of the AI processing part 213 in FIG. The processing circuit 3022 can also perform the function of the learning unit 2132 in FIG. 2 as described above.
可选地,堆叠式CMOS图像传感器芯片还可包括第三基板。存储器和处理电路可分别位于第二基板和/或第三基板上。即,像素阵列部、存储器和处理单路可分别位于不同的基板上。Optionally, the stacked CMOS image sensor chip may further include a third substrate. The memory and processing circuitry may be located on the second substrate and/or the third substrate, respectively. That is, the pixel array unit, the memory, and the processing unit may be respectively located on different substrates.
基于上述构造,本发明的图像传感器21本身能够对目标物的图像执行AI识别处理,从而能够在离线状态下对目标物进行识别。Based on the above configuration, the image sensor 21 of the present invention can itself perform AI recognition processing on the image of the target object, thereby enabling recognition of the target object in an offline state.
[AI模型的创建/训练方法][AI model creation/training method]
图4示出根据本发明的创建或再训练AI模型的步骤。在步骤S01中,AI处理部(具体地,位于云环境或图像传感器中的学习部2132)获取学习数据,该学习数据包括物品图像的数据集以及物品的名称和属性等,并且学习数据还包括AI称重系统的历史识别数据;在步骤S02中,学习部2132利用学习数据创建AI模型,或者对已创建的AI模型进行训练或再训练;在步骤S03中,输出创建的或更新后的AI模型。Fig. 4 illustrates the steps of creating or retraining an AI model according to the present invention. In step S01, the AI processing part (specifically, the learning part 2132 located in the cloud environment or image sensor) acquires learning data, the learning data includes the data set of the item image and the name and attribute of the item, etc., and the learning data also includes historical identification data of the AI weighing system; in step S02, the learning part 2132 uses the learning data to create an AI model, or trains or retrains the created AI model; in step S03, the created or updated AI model is output.
根据本发明,用于创建或训练AI模型的图像数据包括多种不同的图像数据集,例如物品的轮廓图像、带包装的物品图像、物品截面图像、物品的局部图像等。利用上述数据集中两种以上的数据集来创建本发明的AI模型。具体地,在步骤S01中,可收集物品的整体轮廓的多组图像(数据集1),例如不同角度拍摄的物品轮廓;带包装的物品的多组图像(数据集2),例如被装在各种颜色的包装(例如,各种常见的塑料袋)中的物品的图像;物品的不同截面的多组图像(数据集3),例如被分割为不同形状的物品的图像;以及物品的局部放大图像(数据集4)等。对于不同的物品,可选择使用不同的数据集进行建模。例如,对于西瓜等可能会被分割出售的商品,可收集整体图像和分割后的截面图像两种数据集(即,数据集1和3)。因此,在顾客选择购买例如半个西瓜时,能够快速且精确地识别出商品。再如,对于葡萄,可选择使用整体图像和局部图像这两种数据集(即,数据集1和4),通过对局部细节的确认,能够快速识别出葡萄。According to the present invention, the image data used to create or train the AI model includes a variety of different image data sets, such as outline images of items, images of items with packaging, cross-sectional images of items, partial images of items, and the like. The AI model of the present invention is created using more than two kinds of data sets from the above-mentioned data sets. Specifically, in step S01, multiple sets of images of the overall outline of the item (data set 1), such as the profile of the item taken from different angles; multiple sets of images of the packaged item (data set 2), such as images of items in packaging of various colors (for example, various common plastic bags); multiple sets of images of different sections of the item (data set 3), such as images of items segmented into different shapes; and local enlarged images of the item (data set 4). For different items, you can choose to use different data sets for modeling. For example, for watermelon and other commodities that may be sold separately, two data sets (ie, data sets 1 and 3) can be collected for the overall image and the segmented cross-sectional image. Therefore, when a customer chooses to buy, for example, half a watermelon, the product can be quickly and accurately identified. As another example, for grapes, two data sets (namely, data sets 1 and 4) can be selected to use the overall image and the partial image, and the grapes can be quickly identified by confirming the local details.
以上描述的用于创建或训练AI模型的物品图像数据一般为RGB图像或黑白图像。但是,根据本发明优选地,用于创建或训练AI模型的图像数据还可包括物品的三维立体数据、偏光数据、多波长数据中的一种或多种。The item image data used to create or train the AI model described above is generally an RGB image or a black and white image. However, preferably according to the present invention, the image data used for creating or training the AI model may also include one or more of three-dimensional data, polarization data, and multi-wavelength data of the object.
具体地,可利用ToF传感器获得物品的三维立体数据,即ToF数据(数据集5),并且将其与物品的RGB图像(数据集1至4中的至少一个)进行融合,获得物品的立体图像。立体图像能够完整呈现物品的表面特征,从而优化AI模型的识别精度。对于存在光滑表面的物品,在建模时,可使用通过偏光传感器获得的物品的偏光图像(数据集6)。通过使用偏光图像,能够避免物品表面反光而导致拍摄的图像不清楚的问题。另外,还可使用多波长传感器对物品进行拍摄,以获得物品的多波长图像(数据集7)。由于多波长传感器能够捕获物品表面颜色的细小差异,所以多波长图像与物品的RGB图像(数据集1至4中的至少一个)的组合使用,对于用于识别存在许多种类的同种物品(例如,产地不同的相同水果)的建模是有益的。Specifically, the ToF sensor can be used to obtain the three-dimensional data of the item, that is, the ToF data (data set 5), and it is fused with the RGB image of the item (at least one of the data sets 1 to 4) to obtain a stereo image of the item. Stereo images can fully present the surface features of objects, thereby optimizing the recognition accuracy of AI models. For items with smooth surfaces, the polarized image of the item obtained by a polarized light sensor (Dataset 6) can be used when modeling. By using polarized images, it is possible to avoid the problem that the captured image is unclear due to reflections on the surface of the object. In addition, the item can also be photographed using a multi-wavelength sensor to obtain a multi-wavelength image of the item (data set 7). Since multi-wavelength sensors are able to capture small differences in the surface color of items, the combined use of multi-wavelength images with RGB images of items (at least one of datasets 1 to 4) is beneficial for modeling for identifying the presence of many species of the same item (e.g., the same fruit with different origins).
如上所述,通过使用两种以上的上述数据集(数据集1~4),与传统识别模型相比,本发明的AI模型能够显著改善识别精度。并且,通过使建模用的图像数据包括除RGB图像以外的其它图像数据(数据集5~7),本发明进一步地改善了AI模型的识别精度。As described above, by using two or more of the above-mentioned data sets (data sets 1-4), the AI model of the present invention can significantly improve the recognition accuracy compared with the conventional recognition model. Furthermore, the present invention further improves the recognition accuracy of the AI model by making the image data used for modeling include other image data (data sets 5-7) other than RGB images.
另外,在收集物品的整体图像和局部图像时,根据本发明的拍摄装置对物品的局部进行放大拍摄和处理,是基于光学变焦而不是现有技术中普遍采用的数字变焦。相应地,在对目标物进行识别时,根据本发明的拍摄装置对目标物的局部放大图像的获取同样是基于光学变焦。因此,图像质量不会因为放大图像而降低,从而进一步改善识别精度。这一点将在下面参考图6进行详细描述。In addition, when collecting the overall image and partial image of the item, the photographing device according to the present invention zooms in and processes the part of the item based on optical zoom instead of digital zoom commonly used in the prior art. Correspondingly, when identifying the target, the acquisition of the partially enlarged image of the target by the shooting device according to the present invention is also based on optical zoom. Therefore, the image quality is not degraded by enlarging the image, thereby further improving the recognition accuracy. This will be described in detail below with reference to FIG. 6 .
[AI识别方法][AI recognition method]
图5示出利用本发明的AI称重识别系统对目标物进行AI识别的方法的步骤。在步骤S501中,在感应到目标物4被放置在置物台1后,摄像装置2对目标物进行拍摄,获取目标物的多个图像,包括整体轮廓图像、局部图像等,如将在下面根据图6描述的。优选地,摄像装置2还可通过其它传感器获取目标物的除RGB图像以外的其它图像数据,例如ToF数据、偏光图像和多波长图像。在步骤S502中,摄像装置2中的AI处理部213对目标物图像进行AI处理,获得高质量的图像数据,以用于AI识别。这里的AI处理包括从拍摄的图像中截取目标物的局部细节图像,如下面根据图6描述的。优选地,对图像的AI处理还包括将RGB图像与例如ToF传感器等的其他传感器数据进行融合或组合,以优化目标物的图像数据。Fig. 5 shows the steps of the method for AI identification of a target by using the AI weighing identification system of the present invention. In step S501, after sensing that the target object 4 is placed on the storage table 1, the camera 2 takes pictures of the target object and acquires multiple images of the target object, including overall outline images, partial images, etc., as will be described below with reference to FIG. 6 . Preferably, the camera device 2 can also acquire image data other than the RGB image of the target through other sensors, such as ToF data, polarized light images and multi-wavelength images. In step S502, the AI processing unit 213 in the imaging device 2 performs AI processing on the target image to obtain high-quality image data for AI recognition. The AI processing here includes intercepting the local detail image of the target object from the captured image, as described below with reference to FIG. 6 . Preferably, the AI processing of the image further includes fusing or combining the RGB image with other sensor data such as a ToF sensor to optimize the image data of the target object.
图6示出根据本发明的图像传感器的图像输出和/或处理模式。根据本发明的图像传感器21输出的图像的尺寸能够从多种像素尺寸中进行选择。例如,图6中所示的4056×3040像素(12M全传感器尺寸)、1947×1459像素(覆盖整个置物台)或640×480像素(视频图形阵列(VGA)尺寸)。因此,根据控制部的控制,图像传感器21的成像部211可相应选择拍摄全景图像(模式1)、目标物的整体图像(模式2)和局部图像(模式3)。模式3在目标物为例如葡萄时能够优化识别结果。Fig. 6 shows an image output and/or processing mode of an image sensor according to the present invention. The size of the image output by the image sensor 21 according to the present invention can be selected from a variety of pixel sizes. For example, 4056x3040 pixels (12M full sensor size), 1947x1459 pixels (covering the entire shelf) or 640x480 pixels (Video Graphics Array (VGA) size) shown in Figure 6. Therefore, according to the control of the control unit, the imaging unit 211 of the image sensor 21 can choose to capture a panoramic image (mode 1), an overall image of the target object (mode 2) and a partial image (mode 3). Mode 3 can optimize the recognition result when the target object is, for example, grapes.
另一方面,在成像部211输出12M全传感器尺寸图像时,AI处理部213还能够从全传感器尺寸图像中截取目标物区域的图像,并且能够从全传感器尺寸图像或目标物区域图像中截取VGA尺寸的图像。以这种方式,也能够获取目标物的局部细节图像。On the other hand, when the imaging unit 211 outputs a 12M full sensor size image, the AI processing unit 213 can also cut out an image of the object area from the full sensor size image, and can cut out a VGA size image from the full sensor size image or the object area image. In this way, it is also possible to acquire local detailed images of objects.
通过上述方式,本发明能够利用具有AI处理功能的图像传感器输出目标物的局部放大图像。与传统识别技术中使用的数字变焦不同,根据本发明的上述图像放大处理是基于光学变焦进行的,因而放大后的图像质量完全不会受到影响,从而与现有技术相比极大改善识别精度和效率。同时,在这个方面,尽管本发明的摄像装置拍摄的图像的数据量比传统摄像装置的数据量大,但是由于本发明的图像传感器本身具有AI处理和识别功能,而不是如现有技术那样必须将图像数据发送到云端进行处理和识别,因此并不会影响识别速度,同时避免了数据传送过程中的丢失或错误的风险。因此,利用高质量的图像,本发明在不影响识别速度的情况下极大改善了目标物的识别精度。Through the above method, the present invention can utilize an image sensor with an AI processing function to output a partially enlarged image of a target object. Different from the digital zoom used in the traditional recognition technology, the above-mentioned image enlargement process according to the present invention is based on optical zoom, so the quality of the enlarged image will not be affected at all, thereby greatly improving the recognition accuracy and efficiency compared with the prior art. At the same time, in this regard, although the data volume of the image captured by the camera device of the present invention is larger than that of the traditional camera device, since the image sensor of the present invention itself has AI processing and recognition functions, instead of sending image data to the cloud for processing and recognition as in the prior art, the recognition speed will not be affected, and the risk of loss or error during data transmission will be avoided. Therefore, by using high-quality images, the present invention greatly improves the recognition accuracy of the target object without affecting the recognition speed.
在步骤S503中,AI处理部213利用存储器212中的AI推论模型基于在步骤S502中处理后的图像数据对目标物进行识别。识别结果可能包括N(N≥1)个选项。此时,优选地,AI处理部213可结合获得的其它信息对不同选项进行加权,以优化多个选项的排序。其它信息指的是可能会影响用户(例如,超市或卖场内的顾客)对商品的选择的信息,例如识别到的用户的匿名特征数据(如,性别和年龄等)、系统所处位置的地址信息以及天气和温度信息等等。In step S503 , the AI processing unit 213 uses the AI inference model in the memory 212 to recognize the target object based on the image data processed in step S502 . The recognition result may include N (N≥1) options. At this time, preferably, the AI processing unit 213 may weight different options in combination with other obtained information, so as to optimize the ranking of multiple options. Other information refers to information that may affect the user's (for example, a customer in a supermarket or a store)'s choice of goods, such as the identified user's anonymous characteristic data (such as gender and age, etc.), the address information of the system's location, weather and temperature information, etc.
在步骤S504中,AI处理部213将上述识别结果输出到输出部3。输出部3将识别结果显示给用户。如果第1个选项非正确结果,用户可在其它选项中选择正确结果,或者手动输入正确结果。在步骤S505中,AI处理部213将图像数据以及正确的识别结果存储到存储器212中,和/或上传到云环境中的计算设备的存储器中。如前所述,根据反馈机制,利用这些数据可对AI推论模型进行再训练。In step S504 , the AI processing unit 213 outputs the recognition result to the output unit 3 . The output unit 3 displays the recognition result to the user. If the first option is not the correct result, the user can select the correct result in other options, or manually input the correct result. In step S505, the AI processing unit 213 stores the image data and the correct recognition result in the memory 212, and/or uploads them to the memory of the computing device in the cloud environment. As mentioned earlier, this data can be used to retrain the AI inference model according to the feedback mechanism.
<第二实施方式><Second Embodiment>
图7示出根据本发明的AI称重系统的第二实施方式。Fig. 7 shows a second embodiment of the AI weighing system according to the present invention.
根据第一实施方式的拍摄装置2的图像传感器21能够输出RGB彩色图像和/或黑白图像。而在第二实施方式中,如图7所示,摄像装置2还包括诸如ToF传感器22、偏光传感器23和多波长传感器24的其他传感器。这些其他传感器均连接至图像传感器21。AI处理部213能够对各种类型传感器输出的数据进行组合或融合后用于AI识别,以优 化目标物的图像和增加目标物的识别精度。The image sensor 21 of the imaging device 2 according to the first embodiment is capable of outputting RGB color images and/or black and white images. In the second embodiment, however, as shown in FIG. 7 , the imaging device 2 further includes other sensors such as a ToF sensor 22 , a polarization sensor 23 and a multi-wavelength sensor 24 . These other sensors are all connected to the image sensor 21 . The AI processing unit 213 can combine or fuse the data output by various types of sensors for AI recognition, so as to optimize the image of the target object and increase the recognition accuracy of the target object.
ToF传感器22是通过检测光源发出的光被目标物反射后到达传感器的光的飞行时间(时间差),测定到目标物距离的ToF(Time of Flight)式距离图像传感器。AI处理部213能够将从成像部211获得的RGB图像和从ToF传感器22输出的ToF图像进行融合,得到目标物的立体图像。基于该立体图像进行识别能够增加目标物的识别精度。The ToF sensor 22 is a ToF (Time of Flight) distance image sensor that measures the distance to the target by detecting the flight time (time difference) of the light emitted by the light source being reflected by the target and reaching the sensor. The AI processing unit 213 can fuse the RGB image obtained from the imaging unit 211 with the ToF image output from the ToF sensor 22 to obtain a stereoscopic image of the target. Recognition based on the stereo image can increase the recognition accuracy of the target.
偏光传感器23例如是将传统偏光相机中作为独立零部件的偏光元件内置于CMOS图像传感器中获得的传感器。偏光传感器23能拍摄因光反射而看不清的被摄体,例如打包蔬菜的塑料袋会呈现不均匀反光,细腻呈现物体表面的凹凸,从而优化目标物的被摄图像。在目标物由于光滑表面等存在反射时,AI处理部213可选择利用偏光传感器的输出优化目标物的识别精度。The polarization sensor 23 is, for example, a sensor obtained by incorporating a polarization element, which is an independent component in a conventional polarization camera, into a CMOS image sensor. The polarized light sensor 23 can capture objects that cannot be seen clearly due to light reflection. For example, plastic bags for packaging vegetables will show uneven reflection, and the unevenness of the surface of the object will be displayed delicately, thereby optimizing the captured image of the target object. When the target object has reflection due to smooth surface, etc., the AI processing unit 213 can choose to use the output of the polarization sensor to optimize the recognition accuracy of the target object.
多波长传感器24利用对不同波长的光的不同灵敏度能够捕获目标物的颜色的细小差异。利用多波长传感器24输出的目标物对不同波长的灵敏度差异,AI处理部213在识别诸如柑橘类的水果时是特别有利的。The multi-wavelength sensor 24 can capture small differences in the color of the target object by utilizing different sensitivities to different wavelengths of light. The AI processing unit 213 is particularly advantageous when identifying fruits such as citrus by utilizing the difference in sensitivity of the target object to different wavelengths output by the multi-wavelength sensor 24 .
根据本发明的第二实施方式,AI处理部213能够选择性地将上述传感器输出的数据中一种或多种与图像传感器21输出的数据进行组合或融合。如此,与仅使用RGB图像的第一实施方式相比,第二实施方式通过对多种类型的传感器数据的单独或组合使用进一步改善了识别效果。According to the second embodiment of the present invention, the AI processing unit 213 can selectively combine or fuse one or more of the data output by the above sensors with the data output by the image sensor 21 . In this way, compared with the first embodiment using only RGB images, the second embodiment further improves the recognition effect by using multiple types of sensor data individually or in combination.
<第三实施方式><Third Embodiment>
图8示出根据本发明的AI称重系统的第三实施方式。第三实施方式和第一实施方式的不同之处在于包括设置两个摄像装置。如图8所示,根据该实施方式的AI称重系统包括摄像装置2和与摄像装置2通信的摄像装置5。摄像装置2与第一实施方式的相同。Fig. 8 shows a third embodiment of the AI weighing system according to the present invention. The difference between the third embodiment and the first embodiment is that two imaging devices are provided. As shown in FIG. 8 , the AI weighing system according to this embodiment includes a camera 2 and a camera 5 communicating with the camera 2 . The imaging device 2 is the same as that of the first embodiment.
摄像装置5设置在与置物台上的摄像装置2不同的位置处,例如,超市等场所的收银台前。如图8所示,摄像装置5用于拍摄目标人物(例如,顾客)的图像,并且将拍摄的图像发送给摄像装置2。摄像装置2的AI处理部213能够从顾客的图像中获取顾客的匿名个人信息(特征数据),例如表征性别和年龄等的元数据。AI处理部213可仅存储和输出该匿名个人信息,而不会保存顾客的图像。AI处理部213基于与该匿名个人信息关联的偏好信息来辅助识别目标物,例如通过改变识别结果中某一选项的权重。The camera device 5 is arranged at a different position from the camera device 2 on the shelf, for example, in front of the checkout counter of a supermarket or the like. As shown in FIG. 8 , the imaging device 5 is used to capture an image of a target person (for example, a customer), and transmit the captured image to the imaging device 2 . The AI processing unit 213 of the imaging device 2 can acquire anonymous personal information (characteristic data) of the customer, such as metadata representing gender and age, from the image of the customer. The AI processing section 213 may only store and output the anonymous personal information without saving the image of the customer. The AI processing unit 213 assists in identifying the target based on the preference information associated with the anonymous personal information, for example, by changing the weight of an option in the identification result.
另一方面,与摄像装置2类似地,摄像装置5本身也可包括具有AI处理功能的图像传感器。该图像传感器对顾客的图像进行处理,以获得顾客的特征数据(例如,顾客的性别和年龄)。摄像装置5仅将从拍摄的图像中获取的元数据,而不是图像本身,发送给摄像装置2,从而在提高传输和处理速度的同时完全保护个人隐私。On the other hand, similarly to the camera 2, the camera 5 itself may also include an image sensor having an AI processing function. The image sensor processes the image of the customer to obtain characteristic data of the customer (for example, the customer's gender and age). The camera 5 sends only the metadata obtained from the captured image, not the image itself, to the camera 2, thereby fully protecting personal privacy while increasing transmission and processing speed.
优选地,摄像装置5可包括RGB图像传感器和ToF传感器,用于分别获得顾客的RGB图像和ToF图像,并将两种图像发送至摄像装置2的AI处理部213。AI处理部213对顾客的RGB图像和ToF图像进行组合生成顾客的立体图像,并且基于该立体图像获得顾客的特征数据。在摄像装置5本身包括具有AI处理功能的图像传感器的情况下,上述AI处理(包括生成立体图像,和基于该立体图像的特征分析)将在摄像装置5端进行,摄像装置5仅将处理结果发送给摄像装置2。Preferably, the camera device 5 may include an RGB image sensor and a ToF sensor for respectively obtaining the RGB image and the ToF image of the customer, and sending the two images to the AI processing unit 213 of the camera device 2 . The AI processing unit 213 combines the RGB image and the ToF image of the customer to generate a stereoscopic image of the customer, and obtains characteristic data of the customer based on the stereoscopic image. In the case that the camera 5 itself includes an image sensor with an AI processing function, the above-mentioned AI processing (including generating a stereoscopic image, and feature analysis based on the stereoscopic image) will be performed at the camera 5 side, and the camera 5 only sends the processing result to the camera 2.
具体地,举例来说,摄像装置5从拍摄的顾客的图像中获得顾客为25岁左右的女性,并将该元数据(即,“女,25岁”)发送给摄像装置2。摄像装置2将该元数据存储在存储器212中,并发送给推论部2131。存储器212中存储有25岁女性的偏好信息,该偏好信息将根据历史数据定期更新。如图8所示,在将顾客购买的目标物放置在置物台1上后,推论部2131利用AI模型基于成像部211拍摄的目标物图像获得目标物的初步识别结果可能是橘子、橘柑、番茄或柠檬。同时,根据从摄像装置5获得的顾客的信息,推论部2131获知25岁左右的女性更偏好较甜的橘子,并将其作为辅助识别信息。在缺乏其他识别参数时,推论部2131将根据该偏好最终将橘子作为识别结果中第1排序的选项。在顾客支付完成后,上述元数据以及目标物的最终结果(用户是否选择了第一排序的橘子)可用于对AI模型再训练。Specifically, for example, the camera device 5 obtains the customer as a woman about 25 years old from the captured image of the customer, and sends the metadata (ie, "female, 25 years old") to the camera device 2 . The imaging device 2 stores the metadata in the memory 212 and sends it to the inference unit 2131 . The preference information of a 25-year-old female is stored in the memory 212, and the preference information will be regularly updated according to historical data. As shown in FIG. 8, after placing the object purchased by the customer on the storage table 1, the inference unit 2131 uses the AI model to obtain a preliminary recognition result of the object based on the image of the object captured by the imaging unit 211, which may be orange, tangerine, tomato or lemon. At the same time, according to the customer information obtained from the camera 5, the inference unit 2131 learns that women around the age of 25 prefer sweeter oranges, and uses this as auxiliary identification information. In the absence of other recognition parameters, the inference unit 2131 will finally use oranges as the first-ranked option in the recognition results according to the preference. After the customer's payment is completed, the above metadata and the final result of the target object (whether the user selected the first ranked orange) can be used to retrain the AI model.
可选地,摄像装置5还可拍摄顾客持有的购物筐或购物车的图像,并将该图像数据 发送给摄像装置2。摄像装置2可对从摄像装置5获得的图像数据进行预识别。另一方面,可由摄像装置5对其拍摄的图像数据进行用于预识别的AI处理,并直接将处理结果的元数据发送给摄像装置2。例如,摄像装置5可将从图像数据中识别出的物品列表发送给摄像装置2。预识别可作为摄像装置2进行的最终AI识别的辅助信息。Optionally, the camera 5 can also take an image of the shopping basket or shopping cart held by the customer, and send the image data to the camera 2. The camera device 2 can pre-recognize the image data obtained from the camera device 5 . On the other hand, the image data captured by the imaging device 5 can be subjected to AI processing for pre-identification, and the metadata of the processing result can be directly sent to the imaging device 2 . For example, camera device 5 may send a list of objects identified from the image data to camera device 2 . The pre-recognition can be used as auxiliary information for the final AI recognition performed by the imaging device 2 .
并且,在顾客完成支付后,对该顾客的图像的上述AI处理结果的元数据(例如,性别和年龄范围等)以及该顾客购买的商品的最终列表等信息可存储到存储器212和/或云环境上的计算设备中,以用于更新与匿名个人信息关联的偏好信息以及对AI模型的再训练。以此方式,进一步改善AI模型的识别精度。And, after the customer completes the payment, the metadata (for example, gender and age range, etc.) of the above-mentioned AI processing results of the customer's image and the final list of products purchased by the customer can be stored in the storage 212 and/or in the computing device on the cloud environment, so as to update the preference information associated with the anonymous personal information and retrain the AI model. In this way, the recognition accuracy of the AI model is further improved.
在这个方面,在顾客以手机支付时,AI处理部213还可识别顾客的QR码(二维码),获得该顾客的诸如购买历史在内的匿名信息,以用于更新偏好信息和AI模型的再训练。In this regard, when a customer pays with a mobile phone, the AI processing unit 213 can also identify the customer's QR code (two-dimensional code) and obtain anonymous information such as the customer's purchase history for updating preference information and retraining the AI model.
<第四实施方式><Fourth Embodiment>
图9示出根据本发明的AI称重系统的第四实施方式。第四实施方式和第三实施方式的不同之处在于还设置了摄像装置6,即该实施方式包括三个摄像装置2、5和6。摄像装置2和5与第二实施方式的相同。Fig. 9 shows a fourth embodiment of the AI weighing system according to the present invention. The difference between the fourth embodiment and the third embodiment is that an imaging device 6 is also provided, that is, this embodiment includes three imaging devices 2 , 5 and 6 . The imaging devices 2 and 5 are the same as those of the second embodiment.
根据本实施方式的摄像装置6设置在与摄像装置2和5不同的位置处。具体地,多个摄像装置6分别设置在诸如超市等场所中的多个不同位置处。例如,图9示出设置在超市的冷藏柜中的摄像装置6。当然,可在超市的各个商品陈列处都设置摄像装置6。The imaging device 6 according to the present embodiment is provided at a different position from the imaging devices 2 and 5 . Specifically, a plurality of camera devices 6 are respectively installed at a plurality of different positions in a place such as a supermarket. For example, FIG. 9 shows an imaging device 6 installed in a refrigerator in a supermarket. Certainly, the camera device 6 can be set at each commodity display place of the supermarket.
如图9所示,摄像装置6对拍摄范围内的顾客进行拍摄,并将拍摄的图像发送给摄像装置2。摄像装置2可对顾客的图像进行并行定位和绘制(simultaneous localization and mapping,SLAM)处理。具体来说,AI处理部213对图像中的特征点进行分析,确定特定特征点是否相对于另一图像移动一定矢量,并通过合并先后拍摄的多个图像的特征点数据来生成SLAM地图。以这种方式,结合系统的存储器中存储的安装着摄像装置6的货柜处的商品信息,AI处理部213能够确定顾客在设置摄像装置6的货柜处取出或放下的商品信息。As shown in FIG. 9 , the imaging device 6 photographs customers within the imaging range, and sends the captured image to the imaging device 2 . The camera device 2 can perform simultaneous localization and mapping (SLAM) processing on the image of the customer. Specifically, the AI processing unit 213 analyzes the feature points in the image, determines whether a specific feature point moves by a certain vector relative to another image, and generates a SLAM map by combining feature point data of multiple images taken successively. In this way, the AI processing unit 213 can determine the product information that the customer takes out or puts down at the container where the camera device 6 is installed, in combination with the product information stored in the system memory at the container where the camera device 6 is installed.
可选地,摄像装置6还可对拍摄范围内的顾客的购物车或购物筐进行拍摄,并将拍摄的图像发送给摄像装置2。类似地,AI处理部213可对购物车或购物筐的图像进行SLAM处理,从而确定购物车或购物筐中被放入或取走的商品信息。另外可选地,AI处理部213还可对设置在不同位置处的多个摄像装置6拍摄的购物车或购物筐的图像进行对比,通过图像的差异确定购物车或购物筐中被放入或取走的商品信息。Optionally, the camera device 6 can also take pictures of customers' shopping carts or shopping baskets within the shooting range, and send the captured images to the camera device 2 . Similarly, the AI processing unit 213 can perform SLAM processing on the image of the shopping cart or shopping basket, so as to determine the product information put into or taken away from the shopping cart or shopping basket. Optionally, the AI processing unit 213 can also compare the images of shopping carts or shopping baskets captured by multiple camera devices 6 arranged at different locations, and determine the product information put into or taken away from the shopping cart or shopping basket through image differences.
当然,当顾客及其购物车或购物筐都处于摄像装置6的拍摄范围内时,摄像装置6可将顾客及其购物车或购物筐的整体图像发送给摄像装置2,AI处理部213可基于该整体图像确定顾客放入购物车或购物筐或从购物车或购物筐取走的商品信息。Of course, when the customer and his shopping cart or shopping basket are all within the shooting range of the camera 6, the camera 6 can send the overall image of the customer and his shopping cart or shopping basket to the camera 2, and the AI processing unit 213 can determine the product information that the customer puts into the shopping cart or shopping basket or takes away from the shopping cart or shopping basket based on the overall image.
另一方面,与摄像装置5类似地,摄像装置6本身也可包括具有AI处理功能的图像传感器,由该AI图像传感器在摄像装置6上执行上述AI处理,并将处理结果的元数据发送给摄像装置2。例如,顾客在某时刻处从某货柜取出/放下某商品、顾客的购物车或购物筐在某时刻处被放入或取走某商品。On the other hand, similar to the camera 5 , the camera 6 itself may also include an image sensor with an AI processing function, and the AI image sensor performs the aforementioned AI processing on the camera 6 and sends metadata of the processing result to the camera 2 . For example, a customer takes out/puts down a product from a container at a certain time, a customer's shopping cart or basket is put in or taken out a certain product at a certain time.
以这种方式,由于能够获得顾客在卖场内选择的全部或部分商品的精确信息,使得AI处理部213在识别目标物(顾客的选择的商品列表)时的精确度以及识别效率都得到极大改善。In this way, since the accurate information of all or part of the products selected by the customer in the store can be obtained, the accuracy and recognition efficiency of the AI processing unit 213 in recognizing the target (the customer's selected product list) are greatly improved.
此外,与第三实施方式类似地,摄像装置2可对摄像装置6拍摄的顾客的图像进行处理,获得顾客的匿名信息(例如,性别、年龄等)。在摄像装置6本身具有AI图像传感器时,可仅将上述信息的元数据发送给摄像装置2的AI处理部213。同时,AI处理部213可将顾客的元数据和其选择的商品数据关联起来,并存入存储器212中,用于对AI模型的再训练。In addition, similar to the third embodiment, the imaging device 2 can process the image of the customer captured by the imaging device 6 to obtain the customer's anonymous information (for example, gender, age, etc.). When the imaging device 6 itself has an AI image sensor, only the metadata of the above information can be sent to the AI processing unit 213 of the imaging device 2 . At the same time, the AI processing unit 213 can associate the customer's metadata with the commodity data selected by him and store it in the memory 212 for retraining the AI model.
与第三实施方式类似地,在顾客完成支付后,对该顾客的图像的上述AI处理结果的元数据(例如,性别和年龄范围等)以及该顾客购买的商品的最终列表等信息可存储到存储器212和/或云端中,以用于对AI模型的再训练。以此方式,进一步改善AI模型的识别精度。Similar to the third embodiment, after the customer completes the payment, the metadata (for example, gender and age range, etc.) of the above-mentioned AI processing results of the customer's image and the final list of products purchased by the customer can be stored in the memory 212 and/or in the cloud for retraining the AI model. In this way, the recognition accuracy of the AI model is further improved.
<第五实施方式><Fifth Embodiment>
图10示出根据本发明的AI称重系统的第五实施方式。如图10所示,在相同的店铺内可能设置有多个AI称重支付系统,多个系统之间以无线或有线的方式相互连接。从而,多个AI称重系统互相通信,更新店铺的数据库(包括商品信息,顾客的匿名个人信息和购买历史数据等)。基于更新的数据库,对各个AI图像传感器的AI模型进行再训练,从而不断改善AI模型的识别精度。当然,也可在云环境中的计算设备上进行AI模型的训练和再训练,并将更新后的AI模型部署到各个AI称重系统中。Fig. 10 shows a fifth embodiment of the AI weighing system according to the present invention. As shown in Figure 10, multiple AI weighing and payment systems may be installed in the same store, and multiple systems are connected to each other in a wireless or wired manner. Thus, multiple AI weighing systems communicate with each other to update the store's database (including product information, customer's anonymous personal information and purchase history data, etc.). Based on the updated database, the AI model of each AI image sensor is retrained to continuously improve the recognition accuracy of the AI model. Of course, the training and retraining of the AI model can also be performed on the computing device in the cloud environment, and the updated AI model can be deployed to each AI weighing system.
此外,不同店铺(如图10所示的店铺A和店铺B)的数据库信息也可上传到云环境中的计算设备。利用不同店铺的数据库在云环境中的计算设备上建立大数据库,利用该大数据库创建、训练和再训练AI模型,以及更新各店铺的AI模型。利用云环境,不同店铺的AI称重系统的AI处理单元也可实时地通信。In addition, database information of different stores (store A and store B as shown in FIG. 10 ) can also be uploaded to computing devices in the cloud environment. Use the databases of different stores to build a large database on computing devices in the cloud environment, use this large database to create, train and retrain AI models, and update the AI models of each store. Using the cloud environment, the AI processing units of the AI weighing systems in different stores can also communicate in real time.
另外,各店铺可从例如云环境获取其它信息,例如实时温度、区域地址以及当地气候等。利用这些信息,也可辅助改善AI处理单元对目标物(例如,店铺内的商品)的识别精度。In addition, each store can obtain other information, such as real-time temperature, area address, and local climate, etc., from the cloud environment, for example. Utilizing these information can also assist in improving the recognition accuracy of the AI processing unit for the target object (for example, a product in a store).
以上以超市中的商品称重为例对本发明的应用场景进行了描述。但是,本发明显然并不局限于此。本领域技术人员能够理解的是,本发明的AI识别系统还可以应用于任何需要对商品/物品进行识别的场景,而不局限于称重场景。The application scenario of the present invention is described above by taking the weighing of goods in a supermarket as an example. However, the present invention is obviously not limited to this. Those skilled in the art can understand that the AI recognition system of the present invention can also be applied to any scene where commodities/items need to be recognized, and is not limited to weighing scenes.

Claims (20)

  1. 一种AI称重系统,包括:An AI weighing system, including:
    置物台,用于放置目标物,所述置物台能够对所述目标物进行称重;a storage platform for placing a target object, and the storage platform can weigh the target object;
    第一摄像装置,用于对放置在所述置物台上的所述目标物进行拍摄和识别,所述第一摄像装置包括图像传感器,所述图像传感器能够在离线状态下执行用以识别所述目标物的AI处理;以及A first camera device, configured to photograph and identify the target placed on the storage table, the first camera device includes an image sensor, and the image sensor can perform AI processing for identifying the target in an offline state; and
    输出部,用于输出所述目标物的识别结果以及称重结果。The output unit is used to output the identification result and weighing result of the target object.
  2. 如权利要求1所述的AI称重系统,其中,所述图像传感器包括具有第一基板和第二基板的堆叠式CMOS图像传感器芯片,所述第一基板具有将光信号转化为电信号的多个像素,且所述第二基板具有存储器和处理电路,所述存储器存储有AI模型,且所述处理电路通过使用所述AI模型基于所述电信号来执行所述AI处理。The AI weighing system according to claim 1, wherein the image sensor comprises a stacked CMOS image sensor chip having a first substrate having a plurality of pixels converting optical signals into electrical signals, and a second substrate having a memory and a processing circuit, the memory storing an AI model, and the processing circuit performing the AI processing based on the electrical signal by using the AI model.
  3. 如权利要求2所述的AI称重系统,其中,所述AI模型包括第一推论模型。The AI weighing system of claim 2, wherein said AI model comprises a first inference model.
  4. 如权利要求3所述的AI称重系统,其中,所述堆叠式CMOS图像传感器芯片的所述处理电路生成图像数据,并且所述处理电路包括:The AI weighing system of claim 3, wherein the processing circuit of the stacked CMOS image sensor chip generates image data, and the processing circuit comprises:
    学习部,所述学习部基于所述图像数据再训练所述第一推论模型;以及a learning section that retrains the first inference model based on the image data; and
    推论部,所述推论部利用所述第一推论模型对所述目标物进行识别。an inference unit, the inference unit uses the first inference model to identify the target.
  5. 如权利要求3所述的AI称重系统,其还包括位于云环境中的一个或多个计算设备,所述一个或多个计算设备具有相应的处理器和存储器;The AI weighing system of claim 3, further comprising one or more computing devices located in a cloud environment, the one or more computing devices having corresponding processors and memory;
    其中,所述堆叠式CMOS图像传感器芯片的所述处理电路生成图像数据,并且所述图像数据被发送至所述一个或多个计算设备。Wherein the processing circuitry of the stacked CMOS image sensor chip generates image data and the image data is sent to the one or more computing devices.
  6. 如权利要求5所述的AI称重系统,其中,所述一个或多个计算设备基于所述堆叠式CMOS图像传感器芯片生成的所述图像数据创建第二推论模型,并且直接将所述第二推论模型部署到所述堆叠式CMOS图像传感器芯片的所述存储器中,使得所述第一推论模型被更新。The AI weighing system of claim 5, wherein said one or more computing devices create a second inference model based on said image data generated by said stacked CMOS image sensor chip, and directly deploy said second inference model into said memory of said stacked CMOS image sensor chip such that said first inference model is updated.
  7. 如权利要求4或5所述的AI称重系统,其中,所述堆叠式CMOS图像传感器芯片能够选择所述图像数据的尺寸,所述尺寸包括全传感器尺寸和基于所述处理电路的所述AI处理的视频图形阵列(VGA)尺寸。The AI weighing system according to claim 4 or 5, wherein the stacked CMOS image sensor chip is capable of selecting a size of the image data, the size including a full sensor size and a video graphics array (VGA) size based on the AI processing of the processing circuit.
  8. 如权利要求7所述的AI称重系统,其中,当所述堆叠CMOS图像传感器输出具有全传感器尺寸的所述图像数据时,所述AI处理包括从具有所述全传感器尺寸的所述图像数据中截取所述目标物的整体图像或局部图像,所述局部图像包括VGA尺寸图像。The AI weighing system according to claim 7, wherein, when the stacked CMOS image sensor outputs the image data having a full sensor size, the AI processing includes intercepting an overall image or a partial image of the target from the image data having the full sensor size, the partial image including a VGA size image.
  9. 如权利要求1至6中任一项所述的AI称重系统,其中,所述第一摄像装置拍摄的图像数据包括:The AI weighing system according to any one of claims 1 to 6, wherein the image data captured by the first camera device includes:
    所述目标物的轮廓图像;和/或a profile image of the object; and/or
    所述目标物的截面图像;和/或a cross-sectional image of the object; and/or
    所述目标物的包装图像。The package image for the object in question.
  10. 如权利要求1至6中任一项所述的AI称重系统,其中,所述第一摄像装置还包括ToF传感器,并且所述AI处理包括将所述图像传感器输出的RBG数据和所述ToF传感器输出的ToF数据进行组合。The AI weighing system according to any one of claims 1 to 6, wherein the first camera device further includes a ToF sensor, and the AI processing includes combining RBG data output by the image sensor and ToF data output by the ToF sensor.
  11. 如权利要求1至6中任一项所述的AI称重系统,其中,所述第一摄像装置还包括多波长传感器和/或偏光传感器,并且所述AI处理能够利用所述多波长传感器和/或所述偏光传感器的输出优化对所述目标物的识别。The AI weighing system according to any one of claims 1 to 6, wherein the first camera device further includes a multi-wavelength sensor and/or a polarization sensor, and the AI processing can utilize the output of the multi-wavelength sensor and/or the polarization sensor to optimize the identification of the target.
  12. 如权利要求1至6中任一项所述的AI称重系统,其还包括:The AI weighing system according to any one of claims 1 to 6, further comprising:
    第二摄像装置,用于获取目标人物的图像并发送给所述第一摄像装置,并且其中,The second camera device is used to acquire the image of the target person and send it to the first camera device, and wherein,
    所述AI处理包括从所述第二摄像装置获取的所述目标人物的图像获取特征数据并且利用所述特征数据辅助识别所述目标物。The AI processing includes acquiring characteristic data from the image of the target person captured by the second camera device and using the characteristic data to assist in identifying the target object.
  13. 如权利要求12所述的AI称重系统,其中,所述特征数据包括所述目标人物的性别和年龄。The AI weighing system according to claim 12, wherein the feature data includes the gender and age of the target person.
  14. 如权利要求1至6中任一项所述的AI称重系统,其还包括:The AI weighing system according to any one of claims 1 to 6, further comprising:
    第三摄像装置,用于获取目标人物的图像并发送给所述第一摄像装置,并且其中,The third camera device is used to acquire the image of the target person and send it to the first camera device, and wherein,
    所述AI处理包括对所述第三摄像装置获取的所述目标人物的图像进行SLAM处理,并输出处理结果的元数据以辅助识别所述目标物。The AI processing includes performing SLAM processing on the image of the target person captured by the third camera device, and outputting metadata of the processing result to assist in identifying the target object.
  15. 如权利要求14所述的AI称重系统,其中,所述第三摄像装置还获取所述目标人物的购物车或购物筐的图像并发送给所述第一摄像装置,并且所述AI处理包括对所述购物车或购物筐的图像进行SLAM处理,并输出处理结果的元数据以辅助识别所述目标物。The AI weighing system according to claim 14, wherein the third camera device also acquires an image of the shopping cart or shopping basket of the target person and sends it to the first camera device, and the AI processing includes performing SLAM processing on the image of the shopping cart or shopping basket, and outputting metadata of the processing result to assist in identifying the target object.
  16. 如权利要求14所述的AI称重系统,其中,所述第三摄像装置包括定位在所述目标人物的移动范围内的不同位置处的多个图像传感器。The AI weighing system according to claim 14, wherein the third camera device includes a plurality of image sensors positioned at different positions within the moving range of the target person.
  17. 如权利要求1至6中任一项所述的AI称重系统,其中,所述AI处理还包括获取包括环境温度、区域地址和/或气候条件的其它信息,并能够基于所述其它信息辅助识别所述目标物。The AI weighing system according to any one of claims 1 to 6, wherein the AI processing further includes obtaining other information including ambient temperature, area address and/or weather conditions, and can assist in identifying the target based on the other information.
  18. 一种利用多种数据集来增加AI模型的识别精度的方法,包括:A method for increasing the recognition accuracy of an AI model using multiple data sets, including:
    获取物品的图像数据;Obtain the image data of the item;
    利用学习数据创建AI模型,所述学习数据包括所述图像数据和所述物品的名称以及属性;Create an AI model using learning data, the learning data including the image data and the name and attributes of the item;
    应用所述AI模型识别目标物,Apply the AI model to identify the target object,
    其中,所述学习数据包括以下三种数据集中的至少两种:Wherein, the learning data includes at least two of the following three data sets:
    所述物品的轮廓图像;a profile image of the item;
    所述物品的截面图像;a cross-sectional image of the item;
    所述物品的包装图像。The package image for said item.
  19. 如权利要求18所述的方法,其中,所述物品的所述图像数据包括RGB图像数据以及以下数据中的至少一者:The method of claim 18, wherein said image data of said item comprises RGB image data and at least one of:
    ToF数据;ToF data;
    多波长数据;Multi-wavelength data;
    偏光数据。polarized data.
  20. 如权利要求19所述的方法,其中,所述学习数据还包括所述物品的局部图像的数据集。The method of claim 19, wherein said learning data further comprises a data set of partial images of said item.
PCT/CN2023/071665 2022-01-20 2023-01-10 Ai weighing system, and method for improving precision of ai model by using various types of data sets WO2023138447A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210065055.5 2022-01-20
CN202210065055.5A CN116503724A (en) 2022-01-20 2022-01-20 AI weighing system and method for increasing accuracy of AI model using multiple data sets

Publications (1)

Publication Number Publication Date
WO2023138447A1 true WO2023138447A1 (en) 2023-07-27

Family

ID=87318935

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/071665 WO2023138447A1 (en) 2022-01-20 2023-01-10 Ai weighing system, and method for improving precision of ai model by using various types of data sets

Country Status (2)

Country Link
CN (1) CN116503724A (en)
WO (1) WO2023138447A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107767590A (en) * 2016-08-16 2018-03-06 深圳仪普生科技有限公司 Automatic identification commercialization bar code electronic scale and Automatic identification method
CN109508974A (en) * 2018-11-29 2019-03-22 华南理工大学 A kind of shopping accounting system and method based on Fusion Features
CN110705533A (en) * 2019-09-09 2020-01-17 武汉联析医疗技术有限公司 AI recognition and grabbing system for inspection report
US20200151692A1 (en) * 2018-04-18 2020-05-14 Sbot Technologies, Inc. d/b/a Caper Inc. Systems and methods for training data generation for object identification and self-checkout anti-theft
CN111814614A (en) * 2020-06-28 2020-10-23 袁精侠 Intelligent object-identifying electronic scale weighing method and system
CN112307799A (en) * 2019-07-24 2021-02-02 鲁班嫡系机器人(深圳)有限公司 Gesture recognition method, device, system, storage medium and equipment

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107767590A (en) * 2016-08-16 2018-03-06 深圳仪普生科技有限公司 Automatic identification commercialization bar code electronic scale and Automatic identification method
US20200151692A1 (en) * 2018-04-18 2020-05-14 Sbot Technologies, Inc. d/b/a Caper Inc. Systems and methods for training data generation for object identification and self-checkout anti-theft
CN109508974A (en) * 2018-11-29 2019-03-22 华南理工大学 A kind of shopping accounting system and method based on Fusion Features
CN112307799A (en) * 2019-07-24 2021-02-02 鲁班嫡系机器人(深圳)有限公司 Gesture recognition method, device, system, storage medium and equipment
CN110705533A (en) * 2019-09-09 2020-01-17 武汉联析医疗技术有限公司 AI recognition and grabbing system for inspection report
CN111814614A (en) * 2020-06-28 2020-10-23 袁精侠 Intelligent object-identifying electronic scale weighing method and system

Also Published As

Publication number Publication date
CN116503724A (en) 2023-07-28

Similar Documents

Publication Publication Date Title
US11315262B1 (en) Tracking objects in three-dimensional space using calibrated visual cameras and depth cameras
US11783613B1 (en) Recognizing and tracking poses using digital imagery captured from multiple fields of view
US11030442B1 (en) Associating events with actors based on digital imagery
US11861927B1 (en) Generating tracklets from digital imagery
EP3745296A1 (en) Image monitoring-based commodity sensing system and commodity sensing method
US10803292B2 (en) Separation of objects in images from three-dimensional cameras
RU2739542C1 (en) Automatic registration system for a sales outlet
US11284041B1 (en) Associating items with actors based on digital imagery
CN113498530A (en) Object size marking system and method based on local visual information
JP6549558B2 (en) Sales registration device, program and sales registration method
CN110569789B (en) Commodity combined sku identification method and device
US11281888B2 (en) Separation of objects in images from three-dimensional cameras
US20230037427A1 (en) Identifying barcode-to-product mismatches using point of sale devices and overhead cameras
JP6459447B2 (en) Product ordering device, product ordering method, product price output device, product price output method and program
US20240193230A1 (en) System and method for refining an item identification model based on feedback
US20240013530A1 (en) System and method for aggregating metadata for item identification using digital image processing
US20240020973A1 (en) System and method for capturing images for training of an item identification model
WO2023138447A1 (en) Ai weighing system, and method for improving precision of ai model by using various types of data sets
US20230419392A1 (en) Displaying a three-dimensional image of an item to a user of an online concierge system during order fulfillment
US20240029274A1 (en) System and method for detecting a trigger event for identification of an item
US20240020993A1 (en) System and method for item identification using container-based classification
US20240029405A1 (en) System and method for selecting an item from a plurality of identified items by filtering out back images of the items
US20240029390A1 (en) System and method for identifying an item based on an amount of image information in an image of the item
US20240020333A1 (en) System and method for selecting an item from a plurality of identified items based on a similarity value
US20240029277A1 (en) System and method for camera re-calibration based on an updated homography

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23742762

Country of ref document: EP

Kind code of ref document: A1