CN116342985B - Robust feature learning method for dynamic intelligent container - Google Patents

Robust feature learning method for dynamic intelligent container Download PDF

Info

Publication number
CN116342985B
CN116342985B CN202310109875.4A CN202310109875A CN116342985B CN 116342985 B CN116342985 B CN 116342985B CN 202310109875 A CN202310109875 A CN 202310109875A CN 116342985 B CN116342985 B CN 116342985B
Authority
CN
China
Prior art keywords
significant
feature
features
dynamic intelligent
learning method
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310109875.4A
Other languages
Chinese (zh)
Other versions
CN116342985A (en
Inventor
张健
陶泽
贺建飚
任子欣
熊友曹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Central South University
Original Assignee
Central South University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Central South University filed Critical Central South University
Priority to CN202310109875.4A priority Critical patent/CN116342985B/en
Publication of CN116342985A publication Critical patent/CN116342985A/en
Application granted granted Critical
Publication of CN116342985B publication Critical patent/CN116342985B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/778Active pattern-learning, e.g. online learning of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Abstract

The application provides a robust feature learning method for a dynamic intelligent container, which comprises the following steps of S1, generating a first global inhibition mask G according to a first output feature map M; step S2, generating a second global inhibition mask G'; step S3, obtaining a second feature map M' after global inhibition; s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i 'A'; step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K 'A'; and step S6, adopting the global suppression operation implemented in the steps S1-S3 and the local suppression operation in the steps S4-S5 to finally generate a third output characteristic diagram M. The application can dynamically adjust the suppression intensity of the remarkable characteristics, obtain accurate robust characteristic learning expression, and improve the credibility of the dynamic intelligent container retail commodity identification.

Description

Robust feature learning method for dynamic intelligent container
Technical Field
The application relates to the technical field of dynamic intelligent containers and computer vision, in particular to a robust feature learning method for a dynamic intelligent container.
Background
The dynamic intelligent container is used as an important component of an unmanned retail mode, and provides convenience for consumers to enjoy a quick and efficient consumption mode; meanwhile, an entrance is provided for merchants to search low-cost and high-quality operation modes. However, due to the influence of factors such as consumer behavior habits and shopping scenes, retail commodities of the dynamic intelligent container are extremely easy to be shielded in a plurality of different grades in the retail process, the characteristic representation and extraction of the commodities are seriously influenced, and then related functions of the dynamic intelligent container cannot be effectively completed. When the shielding is serious, the background manual guard can only be relied on to watch the video for screening, so that the core advantages of the related technology cannot be exerted, and due to the reasons of fatigue work of staff and the like, the frequency of missed operation and misoperation is high, and the characteristics of retail goods cannot be learned accurately in real time, so that the automatic and intelligent analysis of the dynamic intelligent container becomes an urgent need.
Although computer vision, man-machine interaction and cloud-edge collaboration techniques have all been greatly developed, robustness has become a limiting factor. If the high-quality characteristics of the dynamic intelligent container retail commodity can be obtained through excavation, the manual working pressure is reduced, the working efficiency and the identification accuracy are improved, and the high-reliability, high-efficiency and high-precision dynamic intelligent container consumption experience can be provided for consumers.
To learn the robustness features of generalization, current research is mostly based on deep learning techniques, which are used during the training phase of neural networks: 1) Data enhancement refers to the enhancement of a digital image by digital image transformation (e.g.: randomly rotated, cropped) to generate training data of different distributions, which is simple and efficient to implement, but requires more memory space and training time. 2) Adding regularization noise refers to randomly injecting regularization noise at the input, weight, etc. of the network to prevent the feature detectors from jointly adapting, and improve the robustness of the model. Such as: a binary matrix mask of the same dimension as the full-connected layer weights is created, the activation values of the neuron training process are randomly suppressed, and the structured random discard of the input continuum takes into account the continuity of the input. However, adding regularized noise has the following problems: a) The discarding rule needs to be designed separately, and the quality of the rule not only affects whether important features are lost, but also can prompt the network to ignore the association between the features. b) The discard units (elements or areas) cannot be predicted in advance, and the discard units are randomly selected so that the discard process is blind and irregular. c) The discard process is interrelated, and when the features between the discard units are mutually coupled, even without the discard units themselves, the interrelation between the features iterates to the next level of the network, resulting in a co-adaptation of the feature detectors. Moreover, the regularized noise adding method is generally only suitable for a certain level of the network, cannot be used in a crossing manner at other levels of the network, and has no convenience (such as an hidden layer and a fully connected layer).
In summary, the commodity identification method of the current dynamic intelligent container mostly improves generalization of the commodity identification method by constructing a large-scale data set through time-consuming and labor-consuming manual labeling or using regularization technologies such as data enhancement. Meanwhile, most of such methods for improving product recognition are limited to certain specific subspaces, and cannot generate enhancement examples with irregularity, instability and susceptibility to environmental factors, and feature robustness learning is limited.
Disclosure of Invention
The application aims to provide a robust feature learning method for a dynamic intelligent container, which comprises the following specific technical scheme:
a robust feature learning method for a dynamic intelligent container, comprising the steps of:
step S1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a deep neural network hidden layer to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G';
step S4, dividing the M into a plurality of continuousSmall scale local feature map m i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Represents a column division interval for the M, W represents a width of the M, H represents a height of the M, +.>Representing the number of the local feature graphs after M division;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container.
Optionally, in step S1, the M ε R D×H×W Wherein D represents the dimension of M, H represents the height of M, and W represents theWidth of M.
Optionally, the G is a binary matrix consisting of 0 and 1, and is overlaid on the M.
Optionally, in step S1, the M is an integer multiple of the G.
Optionally, in step S1, the dimensions, the height and the width of the G and the M are the same.
Optionally, in step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum point in each dimension of the M, where a region where the salient feature of the M is located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is any real number;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1];
Use B 1 The ratio of the significant feature of the M to the non-significant feature of the M is blended with the G, specifically, the ratio of the significant feature of the M in the M ranges from 0 to 1, the ratio of the non-significant feature of the M in the M ranges from 0 to 1, and the ratio of the significant feature of the M to the non-significant feature of the M ranges from 0 to positive infinity.
Optionally, in step S4, the basis for dividing the M is set by integrating the data size used by the deep neural network for training retail goods, the height and width of the M.
Optionally, in step S4The s is r and sc The values of (2) are the same.
Optionally, generating the K in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width i
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Is characterized by the significance of (c) and the m i The region of the non-salient features of (2) is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Is characterized by the significance of (c) and the m i Is not significant, generates the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1];
Use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
Optionally, in step S4, the K i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Location of the upper salient featuresPixel units corresponding to the domains;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
The technical scheme of the application has the following beneficial effects:
(1) According to the robust feature learning method for the dynamic intelligent container, the global inhibition operation implemented by the steps S1-S3 and the local inhibition operation implemented by the steps S4-S5 are adopted, the first output feature map M is decoupled from different scales and expressed as the obvious features related to the real target label and the irrelevant non-obvious features, the expression of the obvious features in the training stage is actively inhibited, the association of the non-obvious features which are difficult to learn by the network learning and the real target label is forced, the network has stronger insight on the obvious features and the non-obvious features at the same time, and the operation performance of the dynamic intelligent container retail commodity identification is improved; in addition, the method adopts the combination of the step S6 and the steps S1-S5, comprehensively considers the interpretability of the extracted robust features, interprets the robust feature extraction process from the human cognition angle by combining the class activation diagram, finally generates a third output feature diagram M', can dynamically adjust the suppression intensity of the remarkable features, obtains accurate robust feature learning expression, and improves the reliability of the dynamic intelligent container retail commodity identification.
(2) The application can flexibly insert and use the first output characteristic diagram M in the step S1 in the convolution layer and the full connection layer without modifying the original network structure, thereby improving the robust characteristic learning capability of dynamic intelligent container retail commodity identification in different scenes.
In addition to the objects, features and advantages described above, the present application has other objects, features and advantages. The present application will be described in further detail with reference to the drawings.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this specification, illustrate embodiments of the application and together with the description serve to explain the application. In the drawings:
FIG. 1 is a flow chart of a robust feature learning method for a dynamic intelligent container in accordance with embodiment 1 of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which are derived by a person skilled in the art based on the embodiments of the application, fall within the scope of protection of the application.
Example 1:
referring to FIG. 1, a robust feature learning method for a dynamic intelligent container, comprising the steps of:
s1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a hidden layer of a deep neural network (particularly a ResNet-56 network) to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G'; comparing the front-back difference between the M and the M', wherein the salient features of the maximum value region in the M are all suppressed;
s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Representing a column division interval for the M, the s r and sc The values of the two are the same and are all 4; w represents the width of M, H represents the height of M, < >>The number of the local feature graphs after M division is represented, and the specific value is 16;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container, specifically, the robustness characteristic learning effect of the dynamic intelligent container is optimal when omega is set to be 0.5 by analyzing the observation result of the commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container.
In step S1, the M ε R D×H×W Wherein D represents the dimension of M, in particular 4096, h represents the height of M, in particular 16, and w represents the width of M, in particular 16.
The G is a binary matrix consisting of 0 and 1, and is overlaid on the M.
In step S1, the M is an integer multiple of the G.
In step S1, the dimensions, height and width of G and M are the same.
In step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum value point in each dimension of the M, where a region where the M salient features are located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is 1;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1]Specifically, p is set by analyzing the observation result of the commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container 1 0.3;
use B 1 The ratio of the significant features of the M and the non-significant features of the M is reconciled with the G, in particular, the significant features of the MThe ratio of the non-significant features of the M in the M ranges from 0 to 1, and the ratio of the significant features of the M to the non-significant features of the M ranges from 0 to positive infinity.
In step S4, the basis for dividing the M is set comprehensively by the data size used by the deep neural network training the retail goods, the height and width of the M.
Generating the K in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width (specifically, the height and width are 4) i
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Significant features and the m i The region of non-salient features is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Significant features and the m i Proportion of non-salient features, generating the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1]Setting p by analyzing the observation result of commodity identification precision index in the retail commodity identification experiment of the dynamic intelligent container 2 0.7;
use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
In step S4, for the K i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Pixel units corresponding to the areas where the salient features are located;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
According to the embodiment 1 of the application, the contribution degree of the salient features and the non-salient features reflected in the visualized image of the class activation diagram to the decision of the network execution on the input image can be mined into the non-salient feature region which cannot be focused in the advanced robust feature learning method. Therefore, the embodiment 1 of the application actively suppresses the expression of the salient features in the training stage, forces the association of the non-salient features which are difficult to learn by the network learning and the real target labels, ensures that the network has stronger insight on the salient features and the non-salient features at the same time, and improves the operation performance of the dynamic intelligent container retail commodity identification. The application can improve the retail commodity identification accuracy by 3% on average on the basis of the retail commodity identification network ResNet-56 in the embodiment 1. Specifically, the application realizes that the average identification accuracy of commodities is 96% on 13 types of beverage commodities, namely, assam milk tea, pecam (canned), fenda (canned), red cow (canned), cola (canned), kangshifu green tea, kangshifu ice black tea, farmer mountain spring (500 ml), snow bi (canned) and Yibao (500 ml).
Table 1 results of identification of beverage products on the retail product identification network ResNet-56 of example 1 and prior art, respectively
The above description is only of the preferred embodiments of the present application and is not intended to limit the present application, but various modifications and variations can be made to the present application by those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the protection scope of the present application.

Claims (9)

1. A robust feature learning method for a dynamic intelligent container, comprising the steps of:
step S1, extracting color, outline and texture information of retail goods in a dynamic intelligent container by adopting a deep neural network hidden layer to obtain a first output characteristic diagram M of the hidden layer; generating a first global suppression mask G according to the dimension, the height and the width of the M;
step S2, traversing the area where the significant features of the M are obtained through a maximum value algorithm, obtaining the area where the non-significant features of the M are located, initializing and assigning the G by utilizing the significant features of the M and the area where the non-significant features of the M are located, and generating a second global inhibition mask G' by using Bernoulli distribution to reconcile the proportion of the significant features of the M and the non-significant features of the M with the G;
step S3, uniformly covering the G 'on the M, so as to complete suppression of the salient features on the M, thereby obtaining a second feature map M' after global suppression, where M '=m×g', and M 'represents a product of corresponding matrix elements of the M and G';
s4, dividing the M into a plurality of continuous small-scale local feature graphs M i At said m i Up-generating a local suppression mask K i ';
wherein ,s r represents a line division interval, s, for the M c Represents a column division interval for the M, W represents a width of the M, H represents a height of the M, +.>Representing the number of the local feature graphs after M division;
step S5, the K is processed i ' uniformly cover the m i On, for completing the m i Inhibition of the above significant features, obtaining M K', wherein ,M K ' represent by->Local feature map m i Performing a spliced characteristic diagram after local inhibition;
step S6, the global suppression operation implemented by the steps S1-S3 and the local suppression operation implemented by the steps S4-S5 are adopted, and finally a third output characteristic diagram M 'is generated, wherein M' is represented by an expression (1):
in the expression (1), ω is an inactivation coefficient, ω e 0, 1), and the numerical value thereof is manually adjusted according to the observation experimental result and the characteristic expression condition in the visualized image of the class activation diagram; the experimental result is a retail commodity identification experimental result of the dynamic intelligent container;
in step S2, traversing by using a maximum algorithm to obtain a position L (d, x, y) of a maximum value point in each dimension of the M, where a region where the M salient features are located is represented by taking L (d, x, y) as a center point and making a square with a side length of 3t, that is, L (d, x±t, y±t), where the value of t is any real number;
when the G is initialized and assigned to the position of the region where the M salient features are located, the G value is 0, and the pixel units corresponding to the region where the M salient features are located are discarded;
when the G is initialized and assigned to the position of the area where the non-significant feature of the M is located, the G is valued to be 1, and the pixel unit corresponding to the area where the non-significant feature of the M is located is reserved;
the process of generating the G' is as follows:
G'=G*B 1 ,B 1 ~Bernoulli(p 1 );
wherein ,B1 Indicating a compliance parameter of size p 1 Bernoulli distribution, p 1 ∈[0,1];
Use B 1 The ratio of the significant feature of the M to the non-significant feature of the M is blended with the G, specifically, the ratio of the significant feature of the M in the M ranges from 0 to 1, the ratio of the non-significant feature of the M in the M ranges from 0 to 1, and the ratio of the significant feature of the M to the non-significant feature of the M ranges from 0 to positive infinity.
2. The robust feature learning method for dynamic intelligent containers according to claim 1, characterized in that in step S1, the me R D×H×W Wherein D represents the dimension of M, H represents the height of M, and W represents the width of M.
3. The robust feature learning method for dynamic intelligent containers according to claim 2, wherein G is a binary matrix consisting of 0 and 1 and is overlaid on the M.
4. A robust feature learning method for dynamic smart containers according to claim 3, characterized in that in step S1, the M is an integer multiple of the G.
5. The robust feature learning method for dynamic intelligent containers according to claim 4, characterized in that in step S1, the dimensions, height and width of the G and M are the same.
6. The robust feature learning method for dynamic intelligent containers according to claim 1, wherein in step S4, the basis employed for dividing the M is set comprehensively by the data size used by the deep neural network for training retail goods, the height and width of the M.
7. The robust feature learning method for dynamic intelligent containers according to claim 1, characterized in that in step S4, the S r and sc The values of (2) are the same.
8. The robust feature learning method for dynamic intelligent containers according to any of the claims 1-7, characterized in that the K is generated in step S4 i The process of' is as follows:
first, createIndependently of and from m i Binarization matrix K having the same height and width i
Second, using a maximum algorithm to locate the m by traversal i Is located at the same time as the m i By the m i Is characterized by the significance of (c) and the m i The region of the non-salient features of (2) is opposite to the K i Performing initialization assignment;
finally, the K is distributed by Bernoulli i Reconciling said m i Is characterized by the significance of (c) and the m i Is not significant, generates the K i ' said K i '=K i *B 2 ,B 2 ~Bernoulli(p 2 );
wherein ,B2 Indicating a compliance parameter of size p 2 Bernoulli distribution, p 2 ∈[0,1];
Use B 2 For the K i Reconciling said m i Is characterized by the significance of (c) and the m i The proportion of non-significant features of (a), in particular, the m i Is characterized by the significance of K i In a ratio in the range of 0 to 1, said m i Is not significant in the K i In a ratio in the range of 0 to 1, said m i Is significant in terms of (a) and (m) i The proportion of non-salient features of (a) ranges from 0 to plus infinity.
9. The robust feature learning method for dynamic intelligent containers according to claim 8, characterized in that in step S4, the K is calculated for i The initialization assignment process comprises the following steps:
when the K is to i Initializing a value at said m i At the location of the region of the salient feature of (2), said K i Take a value of 0, indicating that the m is discarded i Pixel units corresponding to the areas where the salient features are located;
when the K is to i Initializing a value at said m i The K is at the position of the region where the non-salient feature is located i A value of 1, indicating that the m is reserved i And a pixel unit corresponding to the region where the non-salient feature is located.
CN202310109875.4A 2023-02-14 2023-02-14 Robust feature learning method for dynamic intelligent container Active CN116342985B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310109875.4A CN116342985B (en) 2023-02-14 2023-02-14 Robust feature learning method for dynamic intelligent container

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310109875.4A CN116342985B (en) 2023-02-14 2023-02-14 Robust feature learning method for dynamic intelligent container

Publications (2)

Publication Number Publication Date
CN116342985A CN116342985A (en) 2023-06-27
CN116342985B true CN116342985B (en) 2023-09-12

Family

ID=86893839

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310109875.4A Active CN116342985B (en) 2023-02-14 2023-02-14 Robust feature learning method for dynamic intelligent container

Country Status (1)

Country Link
CN (1) CN116342985B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113435531A (en) * 2021-07-07 2021-09-24 中国人民解放军国防科技大学 Zero sample image classification method and system, electronic equipment and storage medium
WO2021244079A1 (en) * 2020-06-02 2021-12-09 苏州科技大学 Method for detecting image target in smart home environment
CN114005096A (en) * 2021-11-09 2022-02-01 河北工业大学 Vehicle weight recognition method based on feature enhancement

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110807465B (en) * 2019-11-05 2020-06-30 北京邮电大学 Fine-grained image identification method based on channel loss function

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021244079A1 (en) * 2020-06-02 2021-12-09 苏州科技大学 Method for detecting image target in smart home environment
CN113435531A (en) * 2021-07-07 2021-09-24 中国人民解放军国防科技大学 Zero sample image classification method and system, electronic equipment and storage medium
CN114005096A (en) * 2021-11-09 2022-02-01 河北工业大学 Vehicle weight recognition method based on feature enhancement

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于局部对比和全局稀有度的显著性检测;贺良杰等;《计算机应用研究》;第2832-2840页 *

Also Published As

Publication number Publication date
CN116342985A (en) 2023-06-27

Similar Documents

Publication Publication Date Title
Li et al. Building-a-nets: Robust building extraction from high-resolution remote sensing images with adversarial networks
CN110717896B (en) Plate strip steel surface defect detection method based on significance tag information propagation model
Zhuang et al. Computer vision-based localisation of picking points for automatic litchi harvesting applications towards natural scenarios
Moore et al. “Lattice Cut”-Constructing superpixels using layer constraints
CN102902956B (en) A kind of ground visible cloud image identifying processing method
Grady et al. Fast approximate random walker segmentation using eigenvector precomputation
CN106920243A (en) The ceramic material part method for sequence image segmentation of improved full convolutional neural networks
CN107784661A (en) Substation equipment infrared image classifying identification method based on region-growing method
Li et al. Example-based image colorization via automatic feature selection and fusion
CN112733614B (en) Pest image detection method with similar size enhanced identification
Cui et al. Single image dehazing by latent region‐segmentation based transmission estimation and weighted L1‐norm regularisation
CN107169417A (en) Strengthened based on multinuclear and the RGBD images of conspicuousness fusion cooperate with conspicuousness detection method
CN115272652A (en) Dense object image detection method based on multiple regression and adaptive focus loss
CN110245620A (en) A kind of non-maximization suppressing method based on attention
CN106570885A (en) Background modeling method based on brightness and texture fusion threshold value
CN103399863B (en) Image search method based on the poor characteristic bag of edge direction
Ding et al. Semantic recognition of workpiece using computer vision for shape feature extraction and classification based on learning databases
Nguyen-Quynh et al. Image colorization using the global scene-context style and pixel-wise semantic segmentation
CN103440651B (en) A kind of multi-tag image labeling result fusion method minimized based on order
Yu et al. Mean shift based clustering of neutrosophic domain for unsupervised constructions detection
CN110472639B (en) Target extraction method based on significance prior information
CN113705579A (en) Automatic image annotation method driven by visual saliency
CN116342985B (en) Robust feature learning method for dynamic intelligent container
CN113591850A (en) Two-stage trademark detection method based on computer vision robustness target detection
Zhou et al. Structure-aware multikernel learning for hyperspectral image classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant