CN116124787A - Autonomous multi-angle joint detection method for appearance of non-lambertian body - Google Patents

Autonomous multi-angle joint detection method for appearance of non-lambertian body Download PDF

Info

Publication number
CN116124787A
CN116124787A CN202211604958.2A CN202211604958A CN116124787A CN 116124787 A CN116124787 A CN 116124787A CN 202211604958 A CN202211604958 A CN 202211604958A CN 116124787 A CN116124787 A CN 116124787A
Authority
CN
China
Prior art keywords
product
appearance
learning
detection method
lambertian
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211604958.2A
Other languages
Chinese (zh)
Inventor
王碧
聂宇松
黄伟东
杨书新
吴剑青
曾博文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangxi University of Science and Technology
Original Assignee
Jiangxi University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangxi University of Science and Technology filed Critical Jiangxi University of Science and Technology
Priority to CN202211604958.2A priority Critical patent/CN116124787A/en
Publication of CN116124787A publication Critical patent/CN116124787A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/8851Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/95Investigating the presence of flaws or contamination characterised by the material or shape of the object to be examined
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/8851Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
    • G01N2021/8887Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges based on image processing techniques
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Biochemistry (AREA)
  • General Physics & Mathematics (AREA)
  • Pathology (AREA)
  • Physics & Mathematics (AREA)
  • Immunology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Analytical Chemistry (AREA)
  • Chemical & Material Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Manipulator (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an autonomous multi-angle joint detection method for a non-lambertian appearance, and relates to the technical field of automatic detection. And detecting the appearance of a single product in a fixed detection scene, and obtaining a multi-view planning strategy by a data-driven artificial intelligence method. And inquiring Top-k approximate neighbor samples of the new product image through a learned view planning strategy and a sample set used for learning, and obtaining a corresponding view planning strategy. The weights of k retrieval view planning strategies are adjusted through samples of new products, and the multi-view planning strategy of any new product image is estimated through a weighted sum mode. And adopting a deep reinforcement learning framework to obtain a visual angle planning strategy by continuous trial and error. Knowledge migration is achieved using a new few sample learning method: the LSH is used to retrieve similar samples and the planning strategy is revised by re-weighting. The gradient compensation trace is used for rewarding sparse: the time sequence characteristics of the states and the gradient compensation traces are combined to accelerate the learning process.

Description

Autonomous multi-angle joint detection method for appearance of non-lambertian body
Technical Field
The invention relates to the technical field of automatic detection, in particular to an autonomous multi-angle joint detection method for the appearance of a non-lambertian body.
Background
The quality of the appearance of the product is an important component of the quality of the product, which directly affects the consumer's sense and reduces its commercial value and brand praise. Specifically, the product appearance inspection includes stain, scratch, pit, shallow tumor, edge defect, pattern defect, and the like. The automatic visual inspection (automated visual inspection, AVI) system has the advantages of high inspection accuracy, high inspection speed, low cost and the like, and has been widely used. However, the products to be inspected are complex and changeable, and it is difficult to design a general visual inspection system. Therefore, how to determine the image acquisition view angle and realize knowledge migration under different scenes is a key for constructing a general AVI system.
Disclosure of Invention
The invention aims to provide an autonomous multi-angle joint detection method for the appearance of a non-lambertian body, which is used for planning and collecting a visual angle sequence for a product through deep reinforcement learning so as to realize product quality detection; knowledge migration among different products and different scenes is realized through migration learning and anti-noise learning, and the product appearance detection automation process is facilitated.
The technical aim of the invention is realized by the following technical scheme: an autonomous multi-angle joint detection method for non-lambertian appearance, comprising the steps of:
s1: detecting the appearance of a single product in a fixed detection scene, and obtaining a multi-view planning strategy by a data-driven artificial intelligence method;
s2: inquiring Top-k approximate neighbor samples of a new product image through a learned view planning strategy and a sample set for learning, and obtaining a corresponding view planning strategy;
s3: the weights of k retrieval view planning strategies are adjusted through samples of new products, and the multi-view planning strategy of any new product image is estimated through a weighted sum mode.
The invention is further provided with: in the step S1, a non-lambertian surface slight flaw detection method based on deep reinforcement learning is adopted for detecting the appearance of a single product; automatic visual detection modeling is adopted, and a camera is utilized to capture images of a product to be detected under any angle; preprocessing the captured image of the product to be detected, and inputting the image into a deep neural network model with a convolution layer; the output manipulator is required to rotate the offset angle of the product to be detected; and further controlling the manipulator to rotate the product and capture a new image until judging whether the product has appearance flaws or not.
The invention is further provided with: enhancing automatic visual detection as a Markov decision process model:
M=<δ,A,R,Р,γ>
where δ is the state space, i.e. the sequence of images that the product has captured; a is a decision space, namely the rotation angle of the manipulator; r is a reward function, namely the numerical good or bad degree of the current action, R: delta x A→R; p is the environmental dynamic, i.e. the mapping from the current state to the new state after the motion change, P is delta x A → delta; gamma e 0,1 is the decay factor, i.e. the impact of the current action on future rewards gradually decreases over time interval.
The invention is further provided with: the deep neural network model adopts the combination of a multi-convolution-pooling layer and a dense layer, the node number of an output layer is |A|, and the value of each action is output; let the network model be f, its inputs be state and action, its output be action value, and the learning of the value function be regarded as a regression problem. The objective function is mean square error:
Figure BDA0003998277670000031
s t representing the state at time t, s t+1 A represents the state at the next time from time t t Indicating the action at time t, r t Indicating rewards, θ t Representing current network parameters.
The invention is further provided with: and step S2 and step S3 adopt a depth reinforcement learning cross-scene strategy correction method based on transfer learning, when the appearance of a product is changed, after capturing an image of a new product, similar product detection is realized through a hash group, k product images obtained through retrieval are input into an existing model, so that a responsive planning strategy is obtained, and weights are used for estimating the strategy to be corrected through learning, so that knowledge transfer is completed.
The invention is further provided with: the local sensitive hash is adopted to help search, the local sensitive hash encourages conflict, and two similar points in the original space have similar hash values through mapping, and vice versa; the input of the local sensitive hash is that a new product image X' is input into a dense vector of an original model after being embedded by a convolution-pooling layer; outputting discrete hash values, and searching k similar original product images x according to the hash values of the new product images; the convolution layer is CNN: R m →R n The local sensitive hash is LSH: R n →П d The method comprises the steps of carrying out a first treatment on the surface of the The retrieval of top-k existing samples of high similarity through a locality sensitive hashing is formally described as:
I(X',K)=TopK x∈D (Sim(LSH(CNN(x)),LSH(CNN(X')))),
wherein, sim is pi d ×Π d →[0,1]Is a similarity function.
The invention is further provided with: learning and adjusting weights, wherein the weights consist of k nodes, a softmax normalized neural network layer adopts a three-layer neural network, an activation function is not used for calculating the weights, the input is k similar view planning strategies, and I (x', k) E R |A|×k Representation, output of the weighting y ε R k And if the weight learning network is g, the objective function is as follows:
Figure BDA0003998277670000041
the invention is further provided with: in order to avoid sampling products for multiple times at the same angle in the rotating process of the manipulator, the image sequence acquired by the current product to be detected is used as a state code, and the coding mode combines historical images of the product according to the maximum rotating times to serve as deep neural network input or realizes the coding of the historical image sequence through a neural network model which can be used for time sequence problems such as a long-period memory network and the like;
the motion coding adopts discrete coding, namely, uniformly dividing [0 degree, 360 degrees ] by taking small offset angles as intervals, and adding two decision items of qualification and disability; the reward function adopts sparse rewards, and when the product is correctly detected as being qualified or defective, the reward of +1 is given; when the quality of the product is judged to be wrong, giving-1 rewards; other actions are rewarded with zero.
The invention is further provided with: the state coding adopts a cyclic neural network (RNN)) to code the time sequence characteristics of the state, and the calculation formula is as follows:
h t =sigmoid(m t w h +b h )
m t =tanh(x t w m1 +m t-1 w m2 +b m )
wherein x is t And h t Input and output, m, respectively, in a recurrent neural network architecture 0 By initialising random assignments, w m1 、w m2 And w h Is three matrices (dense layers), updated by back propagation, sigmoid and tanh are activation functions to normalize the output space.
The invention is further provided with: solving the problem of rewarding sparseness by adopting gradient compensation traces, wherein the formula is as follows:
Figure BDA0003998277670000042
Figure BDA0003998277670000043
/>
wherein θ t Is the weight of each layer of network at the time t, e t Is the critical cumulative amount in the qualification trace, alpha t Is the learning rate, gamma t Is a discount factor; the time difference error is noted as:
δ t =r t +γmax a f(s t+1 ,a)-f(s t ,a t )
Figure BDA0003998277670000051
in summary, the invention has the following beneficial effects: 1. in multi-view based work, view planning is designed and the angle of the captured image is fixed for the specific product appearance and detection scene. The method adopts dynamic multi-view planning, and realizes automatic planning of detection view angles through a data-driven artificial intelligence method. 2. Because after key factors influencing visual information acquisition such as appearance of a product to be detected in a detection scene, light source position and the like change, the detection performance of the product to be detected can be obviously reduced. 3. And solving the problem of rewarding sparseness by adopting a gradient compensation trace, and improving the learning efficiency by combining time sequence characteristics of states and the gradient compensation trace acceleration learning process.
Drawings
FIG. 1 is a shallow pit and scratch of a non-lambertian surface;
FIG. 2 is a block diagram of an exemplary automated vision inspection system operation;
FIG. 3 is a schematic representation of a non-lambertian surface mild flaw detection technique based on deep reinforcement learning in accordance with an embodiment of the present invention;
FIG. 4 is a schematic diagram of a deep reinforcement learning cross-scene strategy correction technique route based on transfer learning according to an embodiment of the invention;
FIG. 5 is a technical route of a reward sparse environment learning acceleration method based on gradient compensation traces according to an embodiment of the invention.
Detailed Description
The invention is described in further detail below with reference to fig. 1-5.
Examples: an autonomous multi-angle joint detection method for non-lambertian appearance, as shown in fig. 1-5, comprises the following steps:
s1: detecting the appearance of a single product in a fixed detection scene, and obtaining a multi-view planning strategy by a data-driven artificial intelligence method;
s2: inquiring Top-k approximate neighbor samples of a new product image through a learned view planning strategy and a sample set for learning, and obtaining a corresponding view planning strategy;
s3: the weights of k retrieval view planning strategies are adjusted through samples of new products, and the multi-view planning strategy of any new product image is estimated through a weighted sum mode.
In order to cope with non-lambertian appearance detection of the smooth coating, in a detection scene of a fixed light source, the angle between a product to be detected, the light source and the camera is adjusted through a mechanical arm. On one hand, the difference between the smooth surface such as shallow pits and the like and the main image during reflection is fully utilized; on the other hand, efforts are made to avoid the situation where the high intensity reflection area covers the minute scratches; meanwhile, the problem of false detection caused by light source reflection is reduced by adjusting the relative angle. Reinforcement learning is a data-driven artificial intelligence method that can effectively address sequence decision-making problems. Meanwhile, considering the problem that image feature extraction is difficult, deep reinforcement learning is selected to control the angle sequence of the rotation of the manipulator. The technical route is shown in figure 3.
The detection of the appearance of a single product is realized by adopting a non-lambertian surface slight flaw detection method based on deep reinforcement learning; automatic visual detection modeling is adopted, and a camera is utilized to capture images of a product to be detected under any angle; preprocessing the captured image of the product to be detected, and inputting the image into a deep neural network model with a convolution layer; the output manipulator is required to rotate the offset angle of the product to be detected; and further controlling the manipulator to rotate the product and capture a new image until judging whether the product has appearance flaws or not.
Before a surface detection learning system is built, it is necessary to model the automatic visual detection as a reinforcement learning processable markov decision process:
M=<δ,A,R,Р,γ>
where δ is the state space, i.e. the sequence of images that the product has captured; a is a decision space, namely the rotation angle of the manipulator; r is a reward function, namely the numerical good or bad degree of the current action, R: delta x A→R; p is the environmental dynamic, i.e. the mapping from the current state to the new state after the motion change, P is delta x A → delta; gamma e 0,1 is the decay factor, i.e. the impact of the current action on future rewards gradually decreases over time interval.
In order to avoid sampling products for multiple times at the same angle in the rotating process of the manipulator, the image sequence acquired by the current product to be detected is used as a state code, and the coding mode combines historical images of the product according to the maximum rotating times to serve as deep neural network input or realizes the coding of the historical image sequence through a neural network model which can be used for time sequence problems such as a long-period memory network and the like;
the motion coding adopts discrete coding, namely, uniformly dividing [0 degree, 360 degrees ] by taking small offset angles as intervals, and adding two decision items of qualification and disability; the reward function adopts sparse rewards, and when the product is correctly detected as being qualified or defective, the reward of +1 is given; when the quality of the product is judged to be wrong, giving-1 rewards; other actions are rewarded with zero. In order to ensure the high efficiency of the automatic visual detection system and reduce the rotation times of the mechanical arm during single product detection, an attenuation coefficient gamma=0.99 is designed.
The deep neural network model adopts the combination of a multi-convolution-pooling layer and a dense layer, the node number of an output layer is |A|, and the value of each action is output; let the network model be f, its inputs be state and action, its output be action value, and the learning of the value function be regarded as a regression problem. The objective function is mean square error:
Figure BDA0003998277670000081
s t representing the state at time t, s t+1 A represents the state at the next time from time t t Indicating the action at time t, r t Indicating rewards, θ t Representing current network parameters.
In order to solve the problems such as light source change, camera equipment change, to-be-detected product appearance change and the like in a product detection scene, the mobility of an automatic visual detection system is improved, and knowledge is quickly migrated through migration learning. The technical route is shown in fig. 4. When the appearance of a product is changed, after capturing an image of a new product, similar product detection is realized through a hash group, k product images obtained through retrieval are input into an existing model, so that a responsive planning strategy is obtained, and weights are used for estimating the strategy to be corrected through learning and adjustment, so that knowledge migration is completed.
To quickly and accurately obtain historical images that are close to new product images, a locality sensitive hash (Locality Sensitive Hashing, LSH) is employed to aid in retrieval. The locality sensitive hashing encourages collisions by mapping such that two similar points in the original space have similar hash values and vice versa. Therefore, it can be used to realize large-scale information storage and retrieval. The input of the local sensitive hash is that a new product image X' is input into a dense vector of an original model after being embedded by a convolution-pooling layer; outputting discrete hash values, and searching k similar original product images x according to the hash values of the new product images; the convolution layer is CNN: R m →R n The local sensitive hash is LSH: R n →П d The method comprises the steps of carrying out a first treatment on the surface of the The retrieval of top-k existing samples of high similarity through a locality sensitive hashing is formally described as:
I(X',K)=TopK x∈D (Sim(LSH(CNN(x)),LSH(CNN(X')))),
wherein, sim is pi d ×Π d →[0,1]Is a similarity function.
Learning and adjusting weights, wherein the weights consist of k nodes, a softmax normalized neural network layer adopts a three-layer neural network, an activation function is not used for calculating the weights, the input is k similar view planning strategies, and I (x', k) E R |A|×k Representation, output of the weighting y ε R k And if the weight learning network is g, the objective function is as follows:
Figure BDA0003998277670000091
in order to solve the problems of slow learning caused by overlong detection sequences and rare rewards in the initial learning stage due to continuous rotation angles, the feedback speed of rewards in the learning process is improved through gradient compensation marks. The technical route is shown in fig. 5. Firstly, coding an image sequence acquired after the product rotates through a cyclic neural network. The state with the time sequence characteristic is then used as the input of the regression task, and the neural network is connected as a function of the value. And finally, when the network weight is updated, replacing the update gradient of the original optimizer by using a gradient compensation trace, and accelerating the updating of the value function.
To reduce the likelihood of repeated accesses from only certain angles of view being possible with the current visual image, a recurrent neural network (Recurrent Neural Network, RNN) is employed to encode the timing characteristics of the states. I.e. the historically observed image information will be present in some form in the encoded image after the mapping of the current image. Thereby reducing duplicate detection at the same angle. Compared to gate cycle units and long and short term memory networks, RNNs have lower computational complexity. In the case of a small data volume, the RNN has similar output performance to its variant network. The calculation formula is as follows:
h t =sigmoid(m t w h +b h )
m t =tanh(x t w m1 +m t-1 w m2 +b m )
wherein x is t And h t Input and output, m, respectively, in a recurrent neural network architecture 0 By initialising random assignments, w m1 、w m2 And w h Is three matrices (dense layers), updated by back propagation, sigmoid and tanh are activation functions to normalize the output space. The training value function network is consistent with the method for constructing the depth network model.
The eligibility trace may be described simply as an accumulation of historical gradients. It allows the value of the history state (action) to be updated with the current prize by changing the update rules. Gradient compensation trace is a qualification trace proposed in this application. The gradient compensation trace does not cause the effect of the reward on the status to decay, or even disappear, over time, compared to the existing qualifying trace. Therefore, the gradient compensation trace accords with the specific condition of the project and can be used for solving the rewarding sparse problem. Solving the problem of rewarding sparseness by adopting gradient compensation traces, wherein the formula is as follows:
Figure BDA0003998277670000101
Figure BDA0003998277670000102
wherein θ t Is the weight of each layer of network at the time t, e t Is the critical cumulative amount in the qualification trace, alpha t Is the learning rate, gamma t Is a discount factor; the time difference error is noted as:
δ t =r t +γmax a f(s t+1 ,a)-f(s t ,a t )
Figure BDA0003998277670000103
the present application uses a new visual inspection method to adapt view angle planning: and adopting a deep reinforcement learning framework to obtain a visual angle planning strategy by continuous trial and error. Knowledge migration is achieved using a new few sample learning method: the LSH is used to retrieve similar samples and the planning strategy is revised by re-weighting. The gradient compensation trace is used for rewarding sparse: the time sequence characteristics of the states and the gradient compensation traces are combined to accelerate the learning process.
The present embodiment is only for explanation of the present invention and is not to be construed as limiting the present invention, and modifications to the present embodiment, which may not creatively contribute to the present invention as required by those skilled in the art after reading the present specification, are all protected by patent laws within the scope of claims of the present invention.

Claims (10)

1. The autonomous multi-angle joint detection method for the appearance of the non-lambertian body is characterized by comprising the following steps of: the method comprises the following steps:
s1: detecting the appearance of a single product in a fixed detection scene, and obtaining a multi-view planning strategy by a data-driven artificial intelligence method;
s2: inquiring Top-k approximate neighbor samples of a new product image through a learned view planning strategy and a sample set for learning, and obtaining a corresponding view planning strategy;
s3: the weights of k retrieval view planning strategies are adjusted through samples of new products, and the multi-view planning strategy of any new product image is estimated through a weighted sum mode.
2. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 1, characterized by: in the step S1, a non-lambertian surface slight flaw detection method based on deep reinforcement learning is adopted for detecting the appearance of a single product; automatic visual detection modeling is adopted, and a camera is utilized to capture images of a product to be detected under any angle; preprocessing the captured image of the product to be detected, and inputting the image into a deep neural network model with a convolution layer; the output manipulator is required to rotate the offset angle of the product to be detected; and further controlling the manipulator to rotate the product and capture a new image until judging whether the product has appearance flaws or not.
3. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 2, characterized by: enhancing automatic visual detection as a Markov decision process model:
M=<δ,A,R,Р,γ>
where δ is the state space, i.e. the sequence of images that the product has captured; a is a decision space, namely the rotation angle of the manipulator; r is a reward function, namely the numerical good or bad degree of the current action, R: delta x A→R; p is the environmental dynamic, i.e. the mapping from the current state to the new state after the motion change, P is delta x A → delta; gamma e 0,1 is the decay factor, i.e. the impact of the current action on future rewards gradually decreases over time interval.
4. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 2, characterized by: the deep neural network model adopts the combination of a multi-convolution-pooling layer and a dense layer, the node number of an output layer is |A|, and the value of each action is output; let the network model be f, its inputs be state and action, its output be action value, and the learning of the value function be regarded as a regression problem. The objective function is mean square error:
Figure FDA0003998277660000021
s t representing the state at time t, s t+1 A represents the state at the next time from time t t Indicating the action at time t, r t Indicating rewards, θ t Representing current network parameters.
5. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 1, characterized by: and step S2 and step S3 adopt a depth reinforcement learning cross-scene strategy correction method based on transfer learning, when the appearance of a product is changed, after capturing an image of a new product, similar product detection is realized through a hash group, k product images obtained through retrieval are input into an existing model, so that a responsive planning strategy is obtained, and weights are used for estimating the strategy to be corrected through learning, so that knowledge transfer is completed.
6. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 5, characterized by: the local sensitive hash is adopted to help search, the local sensitive hash encourages conflict, and two similar points in the original space have similar hash values through mapping, and vice versa; the input of the local sensitive hash is that a new product image X' is input into a dense vector of an original model after being embedded by a convolution-pooling layer; outputting discrete hash values, and searching k similar original product images x according to the hash values of the new product images; the convolution layer is CNN: R m →R n The local sensitive hash is LSH: R n →П d The method comprises the steps of carrying out a first treatment on the surface of the Top-k high similarity searches through local sensitive hashThe existing sample formalized description is:
I(X',K)=TopK x∈D (Sim(LSH(CNN(x)),LSH(CNN(X')))),
wherein, sim is pi d ×Π d →[0,1]Is a similarity function.
7. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 5, characterized by: learning and adjusting weights, wherein the weights consist of k nodes, a softmax normalized neural network layer adopts a three-layer neural network, an activation function is not used for calculating the weights, the input is k similar view planning strategies, and I (x', k) E R |A|×k Representation, output of the weighting y ε R k And if the weight learning network is g, the objective function is as follows:
Figure FDA0003998277660000031
8. the autonomous multi-angle joint detection method for non-lambertian appearance of claim 2, characterized by: in order to avoid sampling products for multiple times at the same angle in the rotating process of the manipulator, the image sequence acquired by the current product to be detected is used as a state code, and the coding mode combines historical images of the product according to the maximum rotating times to serve as deep neural network input or realizes the coding of the historical image sequence through a neural network model which can be used for time sequence problems such as a long-period memory network and the like;
the motion coding adopts discrete coding, namely, uniformly dividing [0 degree, 360 degrees ] by taking small offset angles as intervals, and adding two decision items of qualification and disability; the reward function adopts sparse rewards, and when the product is correctly detected as being qualified or defective, the reward of +1 is given; when the quality of the product is judged to be wrong, giving-1 rewards; other actions are rewarded with zero.
9. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 8, characterized by: state coding uses a cyclic neural network (RNN)) to code the timing characteristics of states, and the calculation formula is as follows:
h t =sigmoid(m t w h +b h )
m t =tanh(x t w m1 +m t-1 w m2 +b m )
wherein x is t And h t Input and output, m, respectively, in a recurrent neural network architecture 0 By initialising random assignments, w m1 、w m2 And w h Is three matrices (dense layers), updated by back propagation, sigmoid and tanh are activation functions to normalize the output space.
10. The autonomous multi-angle joint detection method for non-lambertian appearance of claim 9, characterized by: solving the problem of rewarding sparseness by adopting gradient compensation traces, wherein the formula is as follows:
Figure FDA0003998277660000042
e t+1 =γ t-1 e t +a tθ f(h tt )
wherein θ t Is the weight of each layer of network at the time t, e t Is the critical cumulative amount in the qualification trace, alpha t Is the learning rate, gamma t Is a discount factor; the time difference error is noted as:
δ t =r t +γmax a f(s t+1 ,a)-f(s t ,a t )
Figure FDA0003998277660000041
/>
CN202211604958.2A 2022-12-14 2022-12-14 Autonomous multi-angle joint detection method for appearance of non-lambertian body Pending CN116124787A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211604958.2A CN116124787A (en) 2022-12-14 2022-12-14 Autonomous multi-angle joint detection method for appearance of non-lambertian body

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211604958.2A CN116124787A (en) 2022-12-14 2022-12-14 Autonomous multi-angle joint detection method for appearance of non-lambertian body

Publications (1)

Publication Number Publication Date
CN116124787A true CN116124787A (en) 2023-05-16

Family

ID=86296451

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211604958.2A Pending CN116124787A (en) 2022-12-14 2022-12-14 Autonomous multi-angle joint detection method for appearance of non-lambertian body

Country Status (1)

Country Link
CN (1) CN116124787A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117908495A (en) * 2024-03-12 2024-04-19 苏州昊信精密科技有限公司 Flexible high-precision processing system and processing method based on multiple sensors

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117908495A (en) * 2024-03-12 2024-04-19 苏州昊信精密科技有限公司 Flexible high-precision processing system and processing method based on multiple sensors
CN117908495B (en) * 2024-03-12 2024-07-05 苏州昊信精密科技有限公司 Flexible high-precision processing system and processing method based on multiple sensors

Similar Documents

Publication Publication Date Title
CN108334936B (en) Fault prediction method based on migration convolutional neural network
CN109305534B (en) Self-adaptive control method of belt conveyor of coal wharf based on computer vision
CN109523013B (en) Air particulate matter pollution degree estimation method based on shallow convolutional neural network
CN108021947B (en) A kind of layering extreme learning machine target identification method of view-based access control model
CN110598767A (en) SSD convolutional neural network-based underground drainage pipeline defect identification method
CN109934810B (en) Defect classification method based on improved particle swarm wavelet neural network
CN108956614A (en) A kind of pit rope dynamic method for detection fault detection and device based on machine vision
CN111539132B (en) Dynamic load time domain identification method based on convolutional neural network
CN109919934A (en) A kind of liquid crystal display panel defect inspection method based on the study of multi-source domain depth migration
CN115070780B (en) Industrial robot grabbing method and device based on digital twinning and storage medium
CN116124787A (en) Autonomous multi-angle joint detection method for appearance of non-lambertian body
CN113033555B (en) Visual SLAM closed loop detection method based on metric learning
CN116245826A (en) DCGAN-based part surface defect detection method
CN116049937A (en) Cross-domain bridge damage identification method based on deep learning
CN117289652A (en) Numerical control machine tool spindle thermal error modeling method based on multi-universe optimization
CN113689383B (en) Image processing method, device, equipment and storage medium
CN116630751B (en) Trusted target detection method integrating information bottleneck and uncertainty perception
CN117291898A (en) Surface defect detection method, system and equipment
CN113420590A (en) Robot positioning method, device, equipment and medium in weak texture environment
CN116958074A (en) Object surface defect detection method based on yolov7
CN115239034B (en) Method and system for predicting early defects of wind driven generator blade
CN106951888B (en) Relative coordinate constraint method and positioning method of human face characteristic point
Lee et al. Accelerating multi-class defect detection of building façades using knowledge distillation of DCNN-based model
CN117011219A (en) Method, apparatus, device, storage medium and program product for detecting quality of article
CN112288078B (en) Self-learning, small sample learning and migration learning method and system based on impulse neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination