CN116342985B - Robust feature learning method for dynamic intelligent container - Google Patents
Robust feature learning method for dynamic intelligent container Download PDFInfo
- Publication number
- CN116342985B CN116342985B CN202310109875.4A CN202310109875A CN116342985B CN 116342985 B CN116342985 B CN 116342985B CN 202310109875 A CN202310109875 A CN 202310109875A CN 116342985 B CN116342985 B CN 116342985B
- Authority
- CN
- China
- Prior art keywords
- significant
- feature
- features
- dynamic intelligent
- learning method
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 230000001629 suppression Effects 0.000 claims abstract description 23
- 238000010586 diagram Methods 0.000 claims abstract description 17
- 230000005764 inhibitory process Effects 0.000 claims abstract description 15
- 230000008569 process Effects 0.000 claims description 13
- 238000009826 distribution Methods 0.000 claims description 10
- 239000011159 matrix material Substances 0.000 claims description 10
- 238000012549 training Methods 0.000 claims description 9
- 238000013528 artificial neural network Methods 0.000 claims description 7
- 230000004913 activation Effects 0.000 claims description 6
- 230000002779 inactivation Effects 0.000 claims description 3
- 244000269722 Thea sinensis Species 0.000 description 3
- 238000002474 experimental method Methods 0.000 description 3
- 235000013361 beverage Nutrition 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 101150044441 PECAM1 gene Proteins 0.000 description 1
- 235000006468 Thea sinensis Nutrition 0.000 description 1
- 238000009412 basement excavation Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 235000020279 black tea Nutrition 0.000 description 1
- 230000019771 cognition Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 235000009569 green tea Nutrition 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 239000008267 milk Substances 0.000 description 1
- 210000004080 milk Anatomy 0.000 description 1
- 235000013336 milk Nutrition 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 235000013616 tea Nutrition 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/778—Active pattern-learning, e.g. online learning of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/30—Computing systems specially adapted for manufacturing
Abstract
The application provides a robust feature learning method for a dynamic intelligent container, which comprises the following steps of S1, generating a first global inhibition mask G according to a first output feature map M; step S2, generating a second global inhibition mask G'; step S3, obtaining a second feature map M' after global inhibition; s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i 'A'; step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K 'A'; and step S6, adopting the global suppression operation implemented in the steps S1-S3 and the local suppression operation in the steps S4-S5 to finally generate a third output characteristic diagram M. The application can dynamically adjust the suppression intensity of the remarkable characteristics, obtain accurate robust characteristic learning expression, and improve the credibility of the dynamic intelligent container retail commodity identification.
Description
Technical Field
The application relates to the technical field of dynamic intelligent containers and computer vision, in particular to a robust feature learning method for a dynamic intelligent container.
Background
The dynamic intelligent container is used as an important component of an unmanned retail mode, and provides convenience for consumers to enjoy a quick and efficient consumption mode; meanwhile, an entrance is provided for merchants to search low-cost and high-quality operation modes. However, due to the influence of factors such as consumer behavior habits and shopping scenes, retail commodities of the dynamic intelligent container are extremely easy to be shielded in a plurality of different grades in the retail process, the characteristic representation and extraction of the commodities are seriously influenced, and then related functions of the dynamic intelligent container cannot be effectively completed. When the shielding is serious, the background manual guard can only be relied on to watch the video for screening, so that the core advantages of the related technology cannot be exerted, and due to the reasons of fatigue work of staff and the like, the frequency of missed operation and misoperation is high, and the characteristics of retail goods cannot be learned accurately in real time, so that the automatic and intelligent analysis of the dynamic intelligent container becomes an urgent need.
Although computer vision, man-machine interaction and cloud-edge collaboration techniques have all been greatly developed, robustness has become a limiting factor. If the high-quality characteristics of the dynamic intelligent container retail commodity can be obtained through excavation, the manual working pressure is reduced, the working efficiency and the identification accuracy are improved, and the high-reliability, high-efficiency and high-precision dynamic intelligent container consumption experience can be provided for consumers.
To learn the robustness features of generalization, current research is mostly based on deep learning techniques, which are used during the training phase of neural networks: 1) Data enhancement refers to the enhancement of a digital image by digital image transformation (e.g.: randomly rotated, cropped) to generate training data of different distributions, which is simple and efficient to implement, but requires more memory space and training time. 2) Adding regularization noise refers to randomly injecting regularization noise at the input, weight, etc. of the network to prevent the feature detectors from jointly adapting, and improve the robustness of the model. Such as: a binary matrix mask of the same dimension as the full-connected layer weights is created, the activation values of the neuron training process are randomly suppressed, and the structured random discard of the input continuum takes into account the continuity of the input. However, adding regularized noise has the following problems: a) The discarding rule needs to be designed separately, and the quality of the rule not only affects whether important features are lost, but also can prompt the network to ignore the association between the features. b) The discard units (elements or areas) cannot be predicted in advance, and the discard units are randomly selected so that the discard process is blind and irregular. c) The discard process is interrelated, and when the features between the discard units are mutually coupled, even without the discard units themselves, the interrelation between the features iterates to the next level of the network, resulting in a co-adaptation of the feature detectors. Moreover, the regularized noise adding method is generally only suitable for a certain level of the network, cannot be used in a crossing manner at other levels of the network, and has no convenience (such as an hidden layer and a fully connected layer).
In summary, the commodity identification method of the current dynamic intelligent container mostly improves generalization of the commodity identification method by constructing a large-scale data set through time-consuming and labor-consuming manual labeling or using regularization technologies such as data enhancement. Meanwhile, most of such methods for improving product recognition are limited to certain specific subspaces, and cannot generate enhancement examples with irregularity, instability and susceptibility to environmental factors, and feature robustness learning is limited.
Disclosure of Invention
The application aims to provide a robust feature learning method for a dynamic intelligent container, which comprises the following specific technical scheme:
a robust feature learning method for a dynamic intelligent container, comprising the steps of:
step S1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a deep neural network hidden layer to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G';
step S4, dividing the M into a plurality of continuousSmall scale local feature map m i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Represents a column division interval for the M, W represents a width of the M, H represents a height of the M, +.>Representing the number of the local feature graphs after M division;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container.
Optionally, in step S1, the M ε R D×H×W Wherein D represents the dimension of M, H represents the height of M, and W represents theWidth of M.
Optionally, the G is a binary matrix consisting of 0 and 1, and is overlaid on the M.
Optionally, in step S1, the M is an integer multiple of the G.
Optionally, in step S1, the dimensions, the height and the width of the G and the M are the same.
Optionally, in step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum point in each dimension of the M, where a region where the salient feature of the M is located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is any real number;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1];
Use B 1 The ratio of the significant feature of the M to the non-significant feature of the M is blended with the G, specifically, the ratio of the significant feature of the M in the M ranges from 0 to 1, the ratio of the non-significant feature of the M in the M ranges from 0 to 1, and the ratio of the significant feature of the M to the non-significant feature of the M ranges from 0 to positive infinity.
Optionally, in step S4, the basis for dividing the M is set by integrating the data size used by the deep neural network for training retail goods, the height and width of the M.
Optionally, in step S4The s is r and sc The values of (2) are the same.
Optionally, generating the K in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width i ;
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Is characterized by the significance of (c) and the m i The region of the non-salient features of (2) is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Is characterized by the significance of (c) and the m i Is not significant, generates the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1];
Use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
Optionally, in step S4, the K i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Location of the upper salient featuresPixel units corresponding to the domains;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
The technical scheme of the application has the following beneficial effects:
(1) According to the robust feature learning method for the dynamic intelligent container, the global inhibition operation implemented by the steps S1-S3 and the local inhibition operation implemented by the steps S4-S5 are adopted, the first output feature map M is decoupled from different scales and expressed as the obvious features related to the real target label and the irrelevant non-obvious features, the expression of the obvious features in the training stage is actively inhibited, the association of the non-obvious features which are difficult to learn by the network learning and the real target label is forced, the network has stronger insight on the obvious features and the non-obvious features at the same time, and the operation performance of the dynamic intelligent container retail commodity identification is improved; in addition, the method adopts the combination of the step S6 and the steps S1-S5, comprehensively considers the interpretability of the extracted robust features, interprets the robust feature extraction process from the human cognition angle by combining the class activation diagram, finally generates a third output feature diagram M', can dynamically adjust the suppression intensity of the remarkable features, obtains accurate robust feature learning expression, and improves the reliability of the dynamic intelligent container retail commodity identification.
(2) The application can flexibly insert and use the first output characteristic diagram M in the step S1 in the convolution layer and the full connection layer without modifying the original network structure, thereby improving the robust characteristic learning capability of dynamic intelligent container retail commodity identification in different scenes.
In addition to the objects, features and advantages described above, the present application has other objects, features and advantages. The present application will be described in further detail with reference to the drawings.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this specification, illustrate embodiments of the application and together with the description serve to explain the application. In the drawings:
FIG. 1 is a flow chart of a robust feature learning method for a dynamic intelligent container in accordance with embodiment 1 of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which are derived by a person skilled in the art based on the embodiments of the application, fall within the scope of protection of the application.
Example 1:
referring to FIG. 1, a robust feature learning method for a dynamic intelligent container, comprising the steps of:
s1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a hidden layer of a deep neural network (particularly a ResNet-56 network) to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G'; comparing the front-back difference between the M and the M', wherein the salient features of the maximum value region in the M are all suppressed;
s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Representing a column division interval for the M, the s r and sc The values of the two are the same and are all 4; w represents the width of M, H represents the height of M, < >>The number of the local feature graphs after M division is represented, and the specific value is 16;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container, specifically, the robustness characteristic learning effect of the dynamic intelligent container is optimal when omega is set to be 0.5 by analyzing the observation result of the commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container.
In step S1, the M ε R D×H×W Wherein D represents the dimension of M, in particular 4096, h represents the height of M, in particular 16, and w represents the width of M, in particular 16.
The G is a binary matrix consisting of 0 and 1, and is overlaid on the M.
In step S1, the M is an integer multiple of the G.
In step S1, the dimensions, height and width of G and M are the same.
In step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum value point in each dimension of the M, where a region where the M salient features are located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is 1;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1]Specifically, p is set by analyzing the observation result of the commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container 1 0.3;
use B 1 The ratio of the significant features of the M and the non-significant features of the M is reconciled with the G, in particular, the significant features of the MThe ratio of the non-significant features of the M in the M ranges from 0 to 1, and the ratio of the significant features of the M to the non-significant features of the M ranges from 0 to positive infinity.
In step S4, the basis for dividing the M is set comprehensively by the data size used by the deep neural network training the retail goods, the height and width of the M.
Generating the K in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width (specifically, the height and width are 4) i ;
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Significant features and the m i The region of non-salient features is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Significant features and the m i Proportion of non-salient features, generating the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1]Setting p by analyzing the observation result of commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container 2 0.7;
use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
In step S4, for the K i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Pixel units corresponding to the areas where the salient features are located;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
According to the embodiment 1 of the application, the contribution degree of the salient features and the non-salient features reflected in the visualized image of the class activation diagram to the decision of the network execution on the input image can be mined into the non-salient feature region which cannot be focused in the advanced robust feature learning method. Therefore, the embodiment 1 of the application actively suppresses the expression of the salient features in the training stage, forces the association of the non-salient features which are difficult to learn by the network learning and the real target labels, ensures that the network has stronger insight on the salient features and the non-salient features at the same time, and improves the operation performance of the dynamic intelligent container retail commodity identification. The application can improve the retail commodity identification accuracy by 3% on average on the basis of the retail commodity identification network ResNet-56 in the embodiment 1. Specifically, the application realizes that the average identification accuracy of commodities is 96% on 13 types of beverage commodities, namely, assam milk tea, pecam (canned), fenda (canned), red cow (canned), cola (canned), kangshifu green tea, kangshifu ice black tea, farmer mountain spring (500 ml), snow bi (canned) and Yibao (500 ml).
Table 1 results of identification of beverage products on the retail product identification network ResNet-56 of example 1 and prior art, respectively
The above description is only of the preferred embodiments of the present application and is not intended to limit the present application, but various modifications and variations can be made to the present application by those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the protection scope of the present application.
Claims (9)
1. A robust feature learning method for a dynamic intelligent container, comprising the steps of:
step S1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a deep neural network hidden layer to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G';
s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Represents a column division interval for the M, W represents a width of the M, H represents a height of the M, +.>Representing the number of the local feature graphs after M division;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container;
in step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum value point in each dimension of the M, where a region where the M salient features are located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is any real number;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1];
Use B 1 The ratio of the significant feature of the M to the non-significant feature of the M is blended with the G, specifically, the ratio of the significant feature of the M in the M ranges from 0 to 1, the ratio of the non-significant feature of the M in the M ranges from 0 to 1, and the ratio of the significant feature of the M to the non-significant feature of the M ranges from 0 to positive infinity.
2. The robust feature learning method for dynamic intelligent containers according to claim 1, characterized in that in step S1, the me R D×H×W Wherein D represents the dimension of M, H represents the height of M, and W represents the width of M.
3. The robust feature learning method for dynamic intelligent containers according to claim 2, wherein G is a binary matrix consisting of 0 and 1 and is overlaid on the M.
4. A robust feature learning method for dynamic smart containers according to claim 3, characterized in that in step S1, the M is an integer multiple of the G.
5. The robust feature learning method for dynamic intelligent containers according to claim 4, characterized in that in step S1, the dimensions, height and width of the G and M are the same.
6. The robust feature learning method for dynamic intelligent containers according to claim 1, wherein in step S4, the basis employed for dividing the M is set comprehensively by the data size used by the deep neural network for training retail goods, the height and width of the M.
7. The robust feature learning method for dynamic intelligent containers according to claim 1, characterized in that in step S4, the S r and sc The values of (2) are the same.
8. The robust feature learning method for dynamic intelligent containers according to any of the claims 1-7, characterized in that the K is generated in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width i ;
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Is characterized by the significance of (c) and the m i The region of the non-salient features of (2) is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Is characterized by the significance of (c) and the m i Is not significant, generates the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1];
Use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
9. The robust feature learning method for dynamic intelligent containers according to claim 8, characterized in that in step S4, the K is calculated for i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Pixel units corresponding to the areas where the salient features are located;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310109875.4A CN116342985B (en) | 2023-02-14 | 2023-02-14 | Robust feature learning method for dynamic intelligent container |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310109875.4A CN116342985B (en) | 2023-02-14 | 2023-02-14 | Robust feature learning method for dynamic intelligent container |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116342985A CN116342985A (en) | 2023-06-27 |
CN116342985B true CN116342985B (en) | 2023-09-12 |
Family
ID=86893839
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310109875.4A Active CN116342985B (en) | 2023-02-14 | 2023-02-14 | Robust feature learning method for dynamic intelligent container |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116342985B (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113435531A (en) * | 2021-07-07 | 2021-09-24 | 中国人民解放军国防科技大学 | Zero sample image classification method and system, electronic equipment and storage medium |
WO2021244079A1 (en) * | 2020-06-02 | 2021-12-09 | 苏州科技大学 | Method for detecting image target in smart home environment |
CN114005096A (en) * | 2021-11-09 | 2022-02-01 | 河北工业大学 | Vehicle weight recognition method based on feature enhancement |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110807465B (en) * | 2019-11-05 | 2020-06-30 | 北京邮电大学 | Fine-grained image identification method based on channel loss function |
-
2023
- 2023-02-14 CN CN202310109875.4A patent/CN116342985B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021244079A1 (en) * | 2020-06-02 | 2021-12-09 | 苏州科技大学 | Method for detecting image target in smart home environment |
CN113435531A (en) * | 2021-07-07 | 2021-09-24 | 中国人民解放军国防科技大学 | Zero sample image classification method and system, electronic equipment and storage medium |
CN114005096A (en) * | 2021-11-09 | 2022-02-01 | 河北工业大学 | Vehicle weight recognition method based on feature enhancement |
Non-Patent Citations (1)
Title |
---|
基于局部对比和全局稀有度的显著性检测;贺良杰等;《计算机应用研究》;第2832-2840页 * |
Also Published As
Publication number | Publication date |
---|---|
CN116342985A (en) | 2023-06-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Li et al. | Building-a-nets: Robust building extraction from high-resolution remote sensing images with adversarial networks | |
CN110717896B (en) | Plate strip steel surface defect detection method based on significance tag information propagation model | |
Zhuang et al. | Computer vision-based localisation of picking points for automatic litchi harvesting applications towards natural scenarios | |
Moore et al. | “Lattice Cut”-Constructing superpixels using layer constraints | |
CN102902956B (en) | A kind of ground visible cloud image identifying processing method | |
Grady et al. | Fast approximate random walker segmentation using eigenvector precomputation | |
CN106920243A (en) | The ceramic material part method for sequence image segmentation of improved full convolutional neural networks | |
CN107784661A (en) | Substation equipment infrared image classifying identification method based on region-growing method | |
Li et al. | Example-based image colorization via automatic feature selection and fusion | |
CN112733614B (en) | Pest image detection method with similar size enhanced identification | |
Cui et al. | Single image dehazing by latent region‐segmentation based transmission estimation and weighted L1‐norm regularisation | |
CN107169417A (en) | Strengthened based on multinuclear and the RGBD images of conspicuousness fusion cooperate with conspicuousness detection method | |
CN115272652A (en) | Dense object image detection method based on multiple regression and adaptive focus loss | |
CN110245620A (en) | A kind of non-maximization suppressing method based on attention | |
CN106570885A (en) | Background modeling method based on brightness and texture fusion threshold value | |
CN103399863B (en) | Image search method based on the poor characteristic bag of edge direction | |
Ding et al. | Semantic recognition of workpiece using computer vision for shape feature extraction and classification based on learning databases | |
Nguyen-Quynh et al. | Image colorization using the global scene-context style and pixel-wise semantic segmentation | |
CN103440651B (en) | A kind of multi-tag image labeling result fusion method minimized based on order | |
Yu et al. | Mean shift based clustering of neutrosophic domain for unsupervised constructions detection | |
CN110472639B (en) | Target extraction method based on significance prior information | |
CN113705579A (en) | Automatic image annotation method driven by visual saliency | |
CN116342985B (en) | Robust feature learning method for dynamic intelligent container | |
CN113591850A (en) | Two-stage trademark detection method based on computer vision robustness target detection | |
Zhou et al. | Structure-aware multikernel learning for hyperspectral image classification |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |