CN116342985A - Robust feature learning method for dynamic intelligent container - Google Patents
Robust feature learning method for dynamic intelligent container Download PDFInfo
- Publication number
- CN116342985A CN116342985A CN202310109875.4A CN202310109875A CN116342985A CN 116342985 A CN116342985 A CN 116342985A CN 202310109875 A CN202310109875 A CN 202310109875A CN 116342985 A CN116342985 A CN 116342985A
- Authority
- CN
- China
- Prior art keywords
- feature
- significant
- features
- learning method
- dynamic intelligent
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 40
- 230000001629 suppression Effects 0.000 claims abstract description 23
- 238000010586 diagram Methods 0.000 claims abstract description 17
- 230000005764 inhibitory process Effects 0.000 claims abstract description 15
- 230000008569 process Effects 0.000 claims description 13
- 238000009826 distribution Methods 0.000 claims description 11
- 239000011159 matrix material Substances 0.000 claims description 10
- 238000012549 training Methods 0.000 claims description 9
- 238000013528 artificial neural network Methods 0.000 claims description 7
- 230000004913 activation Effects 0.000 claims description 6
- 230000002779 inactivation Effects 0.000 claims description 3
- 230000007786 learning performance Effects 0.000 abstract 1
- 244000269722 Thea sinensis Species 0.000 description 3
- 238000002474 experimental method Methods 0.000 description 3
- 235000013361 beverage Nutrition 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 101150044441 PECAM1 gene Proteins 0.000 description 1
- 235000006468 Thea sinensis Nutrition 0.000 description 1
- 238000009412 basement excavation Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 235000020279 black tea Nutrition 0.000 description 1
- 230000019771 cognition Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 235000009569 green tea Nutrition 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 239000008267 milk Substances 0.000 description 1
- 210000004080 milk Anatomy 0.000 description 1
- 235000013336 milk Nutrition 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 235000013616 tea Nutrition 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/778—Active pattern-learning, e.g. online learning of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/30—Computing systems specially adapted for manufacturing
Abstract
The invention provides a robust feature learning method for a dynamic intelligent container, which comprises the following steps of S1, generating a first global inhibition mask G according to a first output feature map M; step S2, generating a second global inhibition mask G'; step S3, obtaining a second feature map M' after global inhibition; s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i 'A'; step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K 'A'; and step S6, adopting the global suppression operation implemented in the steps S1-S3 and the local suppression operation in the steps S4-S5 to finally generate a third output characteristic diagram M. The invention can dynamically adjust the suppression of the remarkable characteristicsThe strength is improved, accurate robust feature learning performance is obtained, and the reliability of dynamic intelligent container retail commodity identification is improved.
Description
Technical Field
The invention relates to the technical field of dynamic intelligent containers and computer vision, in particular to a robust feature learning method for a dynamic intelligent container.
Background
The dynamic intelligent container is used as an important component of an unmanned retail mode, and provides convenience for consumers to enjoy a quick and efficient consumption mode; meanwhile, an entrance is provided for merchants to search low-cost and high-quality operation modes. However, due to the influence of factors such as consumer behavior habits and shopping scenes, retail commodities of the dynamic intelligent container are extremely easy to be shielded in a plurality of different grades in the retail process, the characteristic representation and extraction of the commodities are seriously influenced, and then related functions of the dynamic intelligent container cannot be effectively completed. When the shielding is serious, the background manual guard can only be relied on to watch the video for screening, so that the core advantages of the related technology cannot be exerted, and due to the reasons of fatigue work of staff and the like, the frequency of missed operation and misoperation is high, and the characteristics of retail goods cannot be learned accurately in real time, so that the automatic and intelligent analysis of the dynamic intelligent container becomes an urgent need.
Although computer vision, man-machine interaction and cloud-edge collaboration techniques have all been greatly developed, robustness has become a limiting factor. If the high-quality characteristics of the dynamic intelligent container retail commodity can be obtained through excavation, the manual working pressure is reduced, the working efficiency and the identification accuracy are improved, and the high-reliability, high-efficiency and high-precision dynamic intelligent container consumption experience can be provided for consumers.
To learn the robustness features of generalization, current research is mostly based on deep learning techniques, which are used during the training phase of neural networks: 1) Data enhancement refers to the enhancement of a digital image by digital image transformation (e.g.: randomly rotated, cropped) to generate training data of different distributions, which is simple and efficient to implement, but requires more memory space and training time. 2) Adding regularization noise refers to randomly injecting regularization noise at the input, weight, etc. of the network to prevent the feature detectors from jointly adapting, and improve the robustness of the model. Such as: a binary matrix mask of the same dimension as the full-connected layer weights is created, the activation values of the neuron training process are randomly suppressed, and the structured random discard of the input continuum takes into account the continuity of the input. However, adding regularized noise has the following problems: a) The discarding rule needs to be designed separately, and the quality of the rule not only affects whether important features are lost, but also can prompt the network to ignore the association between the features. b) The discard units (elements or areas) cannot be predicted in advance, and the discard units are randomly selected so that the discard process is blind and irregular. c) The discard process is interrelated, and when the features between the discard units are mutually coupled, even without the discard units themselves, the interrelation between the features iterates to the next level of the network, resulting in a co-adaptation of the feature detectors. Moreover, the regularized noise adding method is generally only suitable for a certain level of the network, cannot be used in a crossing manner at other levels of the network, and has no convenience (such as an hidden layer and a fully connected layer).
In summary, the commodity identification method of the current dynamic intelligent container mostly improves generalization of the commodity identification method by constructing a large-scale data set through time-consuming and labor-consuming manual labeling or using regularization technologies such as data enhancement. Meanwhile, most of such methods for improving product recognition are limited to certain specific subspaces, and cannot generate enhancement examples with irregularity, instability and susceptibility to environmental factors, and feature robustness learning is limited.
Disclosure of Invention
The invention aims to provide a robust feature learning method for a dynamic intelligent container, which comprises the following specific technical scheme:
a robust feature learning method for a dynamic intelligent container, comprising the steps of:
step S1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a deep neural network hidden layer to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G';
s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Represents a column division interval for the M, W represents a width of the M, H represents a height of the M, +.>Representing the number of the local feature graphs after M division;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container.
Optionally, in step S1, the M ε R D×H×W Wherein D represents the dimension of M, H represents the height of M, and W represents the width of M.
Optionally, the G is a binary matrix consisting of 0 and 1, and is overlaid on the M.
Optionally, in step S1, the M is an integer multiple of the G.
Optionally, in step S1, the dimensions, the height and the width of the G and the M are the same.
Optionally, in step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum point in each dimension of the M, where a region where the salient feature of the M is located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is any real number;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1];
Use B 1 The ratio of the significant feature of the M to the non-significant feature of the M is blended with the G, specifically, the ratio of the significant feature of the M in the M ranges from 0 to 1, the ratio of the non-significant feature of the M in the M ranges from 0 to 1, and the ratio of the significant feature of the M to the non-significant feature of the M ranges from 0 to positive infinity.
Optionally, in step S4, the basis for dividing the M is set by integrating the data size used by the deep neural network for training retail goods, the height and width of the M.
Optionally, in step S4, the S r and sc The values of (2) are the same.
Optionally, generating the K in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width i ;
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Is characterized by the significance of (c) and the m i The region of the non-salient features of (2) is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Is characterized by the significance of (c) and the m i Is not significant, generates the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1];
Use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
Optionally, in step S4, the K i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Pixel units corresponding to the areas where the salient features are located;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
The technical scheme of the invention has the following beneficial effects:
(1) According to the robust feature learning method for the dynamic intelligent container, the global inhibition operation implemented by the steps S1-S3 and the local inhibition operation implemented by the steps S4-S5 are adopted, the first output feature map M is decoupled from different scales and expressed as the obvious features related to the real target label and the irrelevant non-obvious features, the expression of the obvious features in the training stage is actively inhibited, the association of the non-obvious features which are difficult to learn by the network learning and the real target label is forced, the network has stronger insight on the obvious features and the non-obvious features at the same time, and the operation performance of the dynamic intelligent container retail commodity identification is improved; in addition, the method adopts the combination of the step S6 and the steps S1-S5, comprehensively considers the interpretability of the extracted robust features, interprets the robust feature extraction process from the human cognition angle by combining the class activation diagram, finally generates a third output feature diagram M', can dynamically adjust the suppression intensity of the remarkable features, obtains accurate robust feature learning expression, and improves the reliability of the dynamic intelligent container retail commodity identification.
(2) The invention can flexibly insert and use the first output characteristic diagram M in the step S1 in the convolution layer and the full connection layer without modifying the original network structure, thereby improving the robust characteristic learning capability of dynamic intelligent container retail commodity identification in different scenes.
In addition to the objects, features and advantages described above, the present invention has other objects, features and advantages. The present invention will be described in further detail with reference to the drawings.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the invention. In the drawings:
FIG. 1 is a flow chart of a robust feature learning method for a dynamic intelligent container in accordance with embodiment 1 of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which are derived by a person skilled in the art based on the embodiments of the invention, fall within the scope of protection of the invention.
Example 1:
referring to FIG. 1, a robust feature learning method for a dynamic intelligent container, comprising the steps of:
s1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a hidden layer of a deep neural network (particularly a ResNet-56 network) to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G'; comparing the front-back difference between the M and the M', wherein the salient features of the maximum value region in the M are all suppressed;
s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Representing a column division interval for the M, the s r and sc The values of the two are the same and are all 4; w represents the width of M, H represents the height of M, < >>The number of the local feature graphs after M division is represented, and the specific value is 16;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container, specifically, the robustness characteristic learning effect of the dynamic intelligent container is optimal when omega is set to be 0.5 by analyzing the observation result of the commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container.
In step S1, the M ε R D×H×W Wherein D represents the dimension of M, in particular 4096, h represents the height of M, in particular 16, and w represents the width of M, in particular 16.
The G is a binary matrix consisting of 0 and 1, and is overlaid on the M.
In step S1, the M is an integer multiple of the G.
In step S1, the dimensions, height and width of G and M are the same.
In step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum value point in each dimension of the M, where a region where the M salient features are located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is 1;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1]Specifically, p is set by analyzing the observation result of the commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container 1 0.3;
use B 1 The ratio of the significant feature of the M to the non-significant feature of the M is blended with the G, specifically, the ratio of the significant feature of the M in the M ranges from 0 to 1, the ratio of the non-significant feature of the M in the M ranges from 0 to 1, and the ratio of the significant feature of the M to the non-significant feature of the M ranges from 0 to positive infinity.
In step S4, the basis for dividing the M is set comprehensively by the data size used by the deep neural network training the retail goods, the height and width of the M.
Generating the K in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width (specifically, the height and width are 4) i ;
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Significant features and the m i The region of non-salient features is opposite to the K i Performing initialization assignment;
finally, bernoulli distribution is usedFor the K i Reconciling said m i Significant features and the m i Proportion of non-salient features, generating the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1]Setting p by analyzing the observation result of commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container 2 0.7;
use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
In step S4, for the K i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Pixel units corresponding to the areas where the salient features are located;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
According to the embodiment 1 of the invention, the contribution degree of the salient features and the non-salient features reflected in the visualized image of the class activation diagram to the decision of the network execution on the input image can be mined into the non-salient feature region which cannot be focused in the advanced robust feature learning method. Therefore, the embodiment 1 of the invention actively suppresses the expression of the salient features in the training stage, forces the association of the non-salient features which are difficult to learn by the network learning and the real target labels, ensures that the network has stronger insight on the salient features and the non-salient features at the same time, and improves the operation performance of the dynamic intelligent container retail commodity identification. The invention can improve the retail commodity identification accuracy by 3% on average on the basis of the retail commodity identification network ResNet-56 in the embodiment 1. Specifically, the invention realizes that the average identification accuracy of commodities is 96% on 13 types of beverage commodities, namely, assam milk tea, pecam (canned), fenda (canned), red cow (canned), cola (canned), kangshifu green tea, kangshifu ice black tea, farmer mountain spring (500 ml), snow bi (canned) and Yibao (500 ml).
Table 1 results of identification of beverage products on the retail product identification network ResNet-56 of example 1 and prior art, respectively
The above description is only of the preferred embodiments of the present invention and is not intended to limit the present invention, but various modifications and variations can be made to the present invention by those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (10)
1. A robust feature learning method for a dynamic intelligent container, comprising the steps of:
step S1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a deep neural network hidden layer to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G';
s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Represents a column division interval for the M, W represents a width of the M, H represents a height of the M, +.>Representing the number of the local feature graphs after M division;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container.
2. The robust feature learning method for dynamic intelligent containers according to claim 1, characterized in that in step S1, the me R D×H×W Wherein D represents the dimension of M, H represents the height of M, and W represents the width of M.
3. The robust feature learning method for dynamic intelligent containers according to claim 2, wherein G is a binary matrix consisting of 0 and 1 and is overlaid on the M.
4. A robust feature learning method for dynamic smart containers according to claim 3, characterized in that in step S1, the M is an integer multiple of the G.
5. The robust feature learning method for dynamic intelligent containers according to claim 4, characterized in that in step S1, the dimensions, height and width of the G and M are the same.
6. The robust feature learning method for a dynamic intelligent container according to claim 5, wherein in step S2, a maximum value algorithm is used to traverse and obtain a position L (d, x, y) of a maximum value point in each dimension of M, wherein a region where the salient feature of M is located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, namely L (d, x±t, y±t), and the value of t is any real number;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1];
Use B 1 The ratio of the significant feature of the M to the non-significant feature of the M is blended with the G, specifically, the ratio of the significant feature of the M in the M ranges from 0 to 1, the ratio of the non-significant feature of the M in the M ranges from 0 to 1, and the ratio of the significant feature of the M to the non-significant feature of the M ranges from 0 to positive infinity.
7. The robust feature learning method for dynamic intelligent containers according to claim 1, wherein in step S4, the basis employed for dividing the M is set comprehensively by the data size used by the deep neural network for training retail goods, the height and width of the M.
8. The robust feature learning method for dynamic intelligent containers according to claim 1, characterized in that in step S4, the S r and sc The values of (2) are the same.
9. The robust feature learning method for dynamic smart containers according to any of the claims 1-8, characterized in that the K is generated in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width i ;
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Is characterized by the significance of (c) and the m i The region of the non-salient features of (2) is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Is characterized by the significance of (c) and the m i Is not significant, generates the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1];
Use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
10. The robust feature learning method for dynamic intelligent containers according to claim 9, characterized in that in step S4, the K is calculated for i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Image corresponding to the region of the upper salient featureA prime unit;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310109875.4A CN116342985B (en) | 2023-02-14 | 2023-02-14 | Robust feature learning method for dynamic intelligent container |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310109875.4A CN116342985B (en) | 2023-02-14 | 2023-02-14 | Robust feature learning method for dynamic intelligent container |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116342985A true CN116342985A (en) | 2023-06-27 |
CN116342985B CN116342985B (en) | 2023-09-12 |
Family
ID=86893839
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310109875.4A Active CN116342985B (en) | 2023-02-14 | 2023-02-14 | Robust feature learning method for dynamic intelligent container |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116342985B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210133479A1 (en) * | 2019-11-05 | 2021-05-06 | Beijing University Of Posts And Telecommunications | Fine-grained image recognition method, electronic device and storage medium |
CN113435531A (en) * | 2021-07-07 | 2021-09-24 | 中国人民解放军国防科技大学 | Zero sample image classification method and system, electronic equipment and storage medium |
WO2021244079A1 (en) * | 2020-06-02 | 2021-12-09 | 苏州科技大学 | Method for detecting image target in smart home environment |
CN114005096A (en) * | 2021-11-09 | 2022-02-01 | 河北工业大学 | Vehicle weight recognition method based on feature enhancement |
-
2023
- 2023-02-14 CN CN202310109875.4A patent/CN116342985B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210133479A1 (en) * | 2019-11-05 | 2021-05-06 | Beijing University Of Posts And Telecommunications | Fine-grained image recognition method, electronic device and storage medium |
WO2021244079A1 (en) * | 2020-06-02 | 2021-12-09 | 苏州科技大学 | Method for detecting image target in smart home environment |
CN113435531A (en) * | 2021-07-07 | 2021-09-24 | 中国人民解放军国防科技大学 | Zero sample image classification method and system, electronic equipment and storage medium |
CN114005096A (en) * | 2021-11-09 | 2022-02-01 | 河北工业大学 | Vehicle weight recognition method based on feature enhancement |
Non-Patent Citations (2)
Title |
---|
PINGPING ZHANG ET AL.: "Learning Uncertain Convolutional Features for Accurate Saliency Detection", 《2017ICCV》, pages 212 - 221 * |
贺良杰等: "基于局部对比和全局稀有度的显著性检测", 《计算机应用研究》, pages 2832 - 2840 * |
Also Published As
Publication number | Publication date |
---|---|
CN116342985B (en) | 2023-09-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Li et al. | Building-a-nets: Robust building extraction from high-resolution remote sensing images with adversarial networks | |
Moore et al. | “Lattice Cut”-Constructing superpixels using layer constraints | |
CN110717896B (en) | Plate strip steel surface defect detection method based on significance tag information propagation model | |
Zhuang et al. | Computer vision-based localisation of picking points for automatic litchi harvesting applications towards natural scenarios | |
CN102902956B (en) | A kind of ground visible cloud image identifying processing method | |
Grady et al. | Fast approximate random walker segmentation using eigenvector precomputation | |
CN109858487B (en) | Weak supervision semantic segmentation method based on watershed algorithm and image category label | |
CN106780485A (en) | SAR image change detection based on super-pixel segmentation and feature learning | |
CN106683046A (en) | Real-time image splicing method for police unmanned aerial vehicle investigation and evidence obtaining | |
CN107133558B (en) | Infrared pedestrian significance detection method based on probability propagation | |
US11531994B2 (en) | Electronic detection of products and arrangement of products in a display structure, electronic detection of objects and arrangement of objects on and around the display structure, electronic detection of conditions of and around the display structure, and electronic scoring of the detected product and object arrangements and of the detected conditions | |
CN109685806B (en) | Image significance detection method and device | |
Cui et al. | Single image dehazing by latent region‐segmentation based transmission estimation and weighted L1‐norm regularisation | |
CN112733614B (en) | Pest image detection method with similar size enhanced identification | |
CN115272652A (en) | Dense object image detection method based on multiple regression and adaptive focus loss | |
Ding et al. | Semantic recognition of workpiece using computer vision for shape feature extraction and classification based on learning databases | |
Nguyen-Quynh et al. | Image colorization using the global scene-context style and pixel-wise semantic segmentation | |
CN110472639B (en) | Target extraction method based on significance prior information | |
CN105160666B (en) | SAR image change detection based on Non-Stationary Analysis and condition random field | |
Yu et al. | Mean shift based clustering of neutrosophic domain for unsupervised constructions detection | |
CN116342985B (en) | Robust feature learning method for dynamic intelligent container | |
CN113705579A (en) | Automatic image annotation method driven by visual saliency | |
CN113591850A (en) | Two-stage trademark detection method based on computer vision robustness target detection | |
CN109523015A (en) | Image processing method in a kind of neural network | |
CN113343900A (en) | Combined nuclear remote sensing image target detection method based on combination of CNN and superpixel |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |