CN114581769A - Method for identifying houses under construction based on unsupervised clustering - Google Patents

Method for identifying houses under construction based on unsupervised clustering Download PDF

Info

Publication number
CN114581769A
CN114581769A CN202210063160.5A CN202210063160A CN114581769A CN 114581769 A CN114581769 A CN 114581769A CN 202210063160 A CN202210063160 A CN 202210063160A CN 114581769 A CN114581769 A CN 114581769A
Authority
CN
China
Prior art keywords
data
category
building
under construction
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210063160.5A
Other languages
Chinese (zh)
Inventor
胡华浪
韩旭
黄进
李剑波
申克建
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Big Data Development Center Of Ministry Of Agriculture And Rural Areas
Southwest Jiaotong University
Original Assignee
Big Data Development Center Of Ministry Of Agriculture And Rural Areas
Southwest Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Big Data Development Center Of Ministry Of Agriculture And Rural Areas, Southwest Jiaotong University filed Critical Big Data Development Center Of Ministry Of Agriculture And Rural Areas
Priority to CN202210063160.5A priority Critical patent/CN114581769A/en
Publication of CN114581769A publication Critical patent/CN114581769A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention discloses a house under construction identification method based on unsupervised clustering, which relates to the technical field of computer vision and comprises the following steps: acquiring image data of a house under construction through an image acquisition device; manually marking the position of the building under construction in the collected image data, and then cutting the marked building under construction to obtain a building image data set; setting the category number of clusters, and then performing category division on the building image data in a comparative clustering unsupervised mode; each building has a corresponding category after being clustered, and the category of the building data is labeled to a label on the original image; carrying out model training on the building data image dataset marked with the label by adopting a Yolox target detection algorithm; adjusting the confidence threshold of each category according to the prediction effect of each category of the house data in the test set; the method is more beneficial to the training and convergence of the model, keeps high recall rate and improves the detection precision.

Description

Method for identifying houses under construction based on unsupervised clustering
Technical Field
The invention relates to the technical field of computer vision, in particular to a house under construction identification method based on unsupervised clustering.
Background
The method comprises the steps that picture or video stream data are collected in real time through a camera, and then the collected picture or video stream data are processed and analyzed through an artificial intelligence technology, so that real-time monitoring of rural houses under construction is realized, and information of the rural houses is pushed to an artificial intelligence platform, so that the houses under construction are judged, and the purpose of intelligent supervision is achieved; the method mainly adopts a target detection technology, at present, the target detection generally adopts a supervised learning mode, namely, a training data set is manually marked, the position and the category of a target are mainly marked, then, a model is trained on the marked data set, and finally, the data is predicted through the trained model; the target detection is one of the most common technologies of artificial intelligence, and is mainly divided into a traditional algorithm adopting manual design characteristics and a detection algorithm based on deep learning, and with the development of hardware technologies such as a deep network and a GPU (graphics processing unit), the target detection performance based on the deep learning is far beyond the traditional algorithm and becomes a mainstream algorithm in the current detection field; the method based on deep learning is also divided into Two types, namely One-Stage (One-Stage) and Two-Stage (Two-Stage), wherein the Two-Stage algorithm mainly represents RCNN series (Rcnn, Fast Rcnn and Fast Rcnn), and the Two-Stage algorithm also means that the Two-Stage algorithm needs to be carried out in Two steps, firstly, a candidate region is predicted, then, the candidate region is classified, and the position of a candidate frame is regressed, so the algorithm is high in precision, but poor in real-time performance; on the contrary, the One-Stage algorithm directly predicts the position and the category of a target in an image and realizes end-to-end detection, so that the algorithm has high real-time performance and is easy to optimize, but the accuracy is reduced compared with a One-Stage algorithm, and in terms of the industry, the algorithm prediction precision is only considered, and the real-time performance is also One of important indexes selected by the algorithm, so that the One-Stage algorithm is the detection algorithm preferred by the industry, and the One-Stage algorithm mainly comprises SSD, RetinaNet and Yolo series algorithms; the Yolo series algorithm focusing on the weight of precision and real-time property is widely applied in the industry from Yolov1 to Yolov5, and then to the Yolo detection algorithm just proposed in this year, the Yolo series algorithm is continuously improved, the performance of the Yolo series algorithm is better and better, and the Yolo algorithm leads the detection precision and the inference speed to be improved by introducing the techniques of Decoupled head, anchor-free, simOTA and the like into the algorithm, so that the Yolo series algorithm becomes one of the most excellent detection algorithms at present.
The main problems and defects of the existing technology are that:
most of the existing target detection algorithms label images manually, but for data of houses under construction, states of houses in different periods are different in the construction process, and because angles shot by cameras are different, the house under construction is different, and meanwhile, the house construction types are different; if a single category (such as 'building houses in place') is adopted to label data, the detection precision is low due to the fact that the difference between houses is large and the model is difficult to converge; meanwhile, all states are summarized into a category, so that the model is not easy to optimize, for example, the data identification is not good if the model is not easy to judge, and the state with poor identification cannot be specially optimized; if adopt the manual work to go to classify to data, because the house data diversity of gathering in the real scene, the angle of shooting including each camera is different, shelters from each other between the house, each state and the diversification of house type during the house construction period, this just causes the categorised difficulty of manual definition, if adopt the good classification of manual definition simultaneously, also need go to carry out manual classification when increasing new training data, it is wrong with data classification easily, cause the model training effect not good.
Disclosure of Invention
The invention aims to: in order to solve the technical problem, the invention provides a house under construction identification method based on unsupervised clustering.
The invention specifically adopts the following technical scheme for realizing the purpose:
the invention relates to a house under construction identification method based on unsupervised clustering, which comprises the following steps:
s1, data set production: acquiring image data of a house under construction through an image acquisition device;
s2, preprocessing of the data set: manually marking the position of the building under construction in the collected image data, and then cutting the marked building under construction to obtain a building image data set;
s3, unsupervised clustering of data: setting the number of the types of the clusters, and then performing type division on the image data of the house under construction obtained in the step S2 in a comparison clustering unsupervised mode;
s4, label redefinition: each building has a corresponding category after being clustered, and the category of the building data is labeled to a label on the original image;
s5, training a detection model: carrying out model training on the building data image dataset marked with the label by adopting a Yolox target detection algorithm;
s6, setting a model confidence threshold: and adjusting the confidence threshold of each category according to the predicted effect of each category of the house data in the test set.
Further, in step S1, image data of different focal lengths at different angles and times of the building under construction are collected by the image collecting device.
Further, in step S2, the labelinmage tool is used to manually label the position of the building in the image.
Further, the step S3 of performing category classification on the building image data in an unsupervised manner of contrast clustering includes the following steps:
under-construction house image data sample xiUsing two data enhancement modes TaAnd TbEnhanced images
Figure RE-GDA0003589141400000031
And
Figure RE-GDA0003589141400000032
comprises the following steps:
Figure RE-GDA0003589141400000033
Figure RE-GDA0003589141400000034
and adopting Resnet34 as a feature extraction network for the enhanced image, wherein the extracted feature vector is expressed as:
Figure RE-GDA0003589141400000035
Figure RE-GDA0003589141400000036
wherein
Figure RE-GDA0003589141400000037
And
Figure RE-GDA0003589141400000038
representing a feature vector, f (-) represents a feature extraction network;
feature vector
Figure RE-GDA0003589141400000039
And
Figure RE-GDA00035891414000000310
and then mapping by two layers of nonlinear perceptrons (MLPs) as follows:
Figure RE-GDA00035891414000000311
Figure RE-GDA00035891414000000312
wherein
Figure RE-GDA00035891414000000313
And
Figure RE-GDA00035891414000000314
representing the mapped feature vector, g (-) represents the nonlinear sensor;
defining samples
Figure RE-GDA00035891414000000315
Loss of
Figure RE-GDA00035891414000000316
Comprises the following steps:
Figure RE-GDA0003589141400000041
Figure RE-GDA0003589141400000042
wherein tau isIIs a temperature parameter of an example level, s (·) represents cosine similarity operation, N represents a set size of the batch, so that after data enhancement, the total number of the batch is 2N, 1 positive sample pair and 2N-2 negative sample pairs can be formed for each sample,
Figure RE-GDA0003589141400000043
and
Figure RE-GDA0003589141400000044
respectively represents that the jth sample in one batch passes through TaAnd TbExtracting feature vectors after data enhancement;
total loss of samples at the instance level linsAs shown in the following formula:
Figure RE-GDA0003589141400000045
when the data samples are mapped to the same number of spatial dimensions as the cluster, each dimension of the data features can be regarded as the probability that the sample belongs to the class, the feature vector extracted by the feature extraction network is mapped into M through two layers of nonlinear perceptors, and M represents the cluster class dimension vector, namely:
Figure RE-GDA0003589141400000046
Figure RE-GDA0003589141400000047
wherein g isc(. cndot.) is a non-linear sensor,
Figure RE-GDA0003589141400000048
and
Figure RE-GDA0003589141400000049
respectively representing the M-dimensional vectors after being mapped by the perceptron, the samples in one batch are at TaAnd TbData enhanced down-composition Ya∈RN×M,Yb∈RN×MA matrix;
definition of
Figure RE-GDA00035891414000000410
Is YaIn the ith column, the first column, similarly,
Figure RE-GDA00035891414000000411
is YbColumn i, therefore sample
Figure RE-GDA00035891414000000412
And
Figure RE-GDA00035891414000000413
loss at cluster level
Figure RE-GDA00035891414000000414
Is defined as:
Figure RE-GDA00035891414000000415
Figure RE-GDA00035891414000000416
wherein tau isCIs a temperture parameter of a cluster level, M represents the number of cluster categories,
Figure RE-GDA0003589141400000051
for the sample at TaData enhances the cluster allocation for the jth category, and, similarly,
Figure RE-GDA0003589141400000052
indicating that the sample is at TbData-enhanced clustering assignment of the jth category;
total loss of samples at cluster level lcluIs represented by the following formula:
Figure RE-GDA0003589141400000053
h (y) assigns information entropy of probability to the clusters, so as to avoid that the network assigns all instances to one cluster, which is specifically expressed as the following formula:
Figure RE-GDA0003589141400000054
Figure RE-GDA0003589141400000055
wherein the content of the first and second substances,
Figure RE-GDA0003589141400000056
is shown at TkThe data of (1) enhances the probability that the jth sample belongs to the category i, | · | | computationally1Represents the norm L1, YkIndicates that a sample in batch is at TkAn output matrix under data enhancement;
the final model loss function is therefore l ═ lins+lclu
And (3) minimizing a loss function through error back propagation to obtain a clustering model, and then classifying the data set through the clustering model.
Further, the data enhancement mode is randomly selected from random clipping, gray scale transformation, chrominance transformation, gaussian blurring and horizontal inversion.
The invention has the following beneficial effects:
aiming at data with complex scenes and a plurality of different states in a building, the invention provides a method for labeling the data in multiple classes in an unsupervised clustering mode, compared with the method for labeling the data in a single class, the data distribution is easier to learn by the multiple classes in the multiple classes mode, so that the convergence fitting is better, and the model is easier to optimize after being divided into a plurality of classes; compared with the condition of manually dividing the building into a plurality of types of labels, the condition of each period of the building under construction is abnormal, complicated and difficult to define along with the conditions of camera shooting angle, house type, house shielding and the like, the potential features among the houses are learned in an unsupervised clustering mode, and the images with high similarity are divided into one type, so that the division of a data set is more accurate, the training and convergence of a model are more facilitated, and the prediction precision of the model on the building under construction is improved; aiming at the condition that the building is divided into a plurality of categories, the prediction precision of each category in the actual scene is different, different threshold values are set for each category according to the prediction condition on the test set, and then the labels of the building are uniformly output, so that the high recall rate can be effectively maintained, and the detection precision can be effectively improved.
Drawings
FIG. 1 is a schematic flow chart of the algorithm provided by the present invention;
FIG. 2 is a schematic diagram of a premise data processing module provided by the present invention;
FIG. 3 is a schematic diagram of a comparison clustering module provided by the present invention;
FIG. 4 is a schematic diagram of the diversification of house data provided by the present invention;
FIG. 5 is a diagram of the detection effect of the present invention using manual classification;
FIG. 6 is a diagram of the detection effect of unsupervised clustering according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. The components of embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations.
Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1
As shown in fig. 1 and fig. 2, the present embodiment provides a house under construction identification method based on unsupervised clustering, including the following steps:
s1, data set production: acquiring image data of a house under construction through an image acquisition device;
in this embodiment, preferably, the image data of each time period, each angle, and different focal lengths of the building under construction are acquired by the image acquisition device. The image acquisition device is a camera, the high-definition camera is controlled by the software platform to acquire all image data of the building under construction in a certain area as acquisition data, and when the data are acquired, data of different scenes and different house types as shown in fig. 4 are collected as much as possible to establish a complete building under construction image data set.
S2, preprocessing of the data set: manually marking the position of the building under construction in the collected image data, and then cutting the marked building under construction to obtain a building image data set;
in this embodiment, preferably, the LabelImage tool is used to manually label the position of the building in the image. And only paying attention to the position information of the house in the image, analyzing and extracting the position coordinates of the house in the image by the marked Xml file, and cutting to obtain a cut image data set of the house in the building.
It should be noted that, collected image data may be labeled only at the location of the building, or at the same time, the category of the building may be labeled, the category of the building may be randomly specified, and when the category is labeled, after the unsupervised way of contrast clustering is adopted for category division, the obtained category is substituted for the labeled category.
S3, unsupervised clustering of data: setting the number of the types of the clusters, and then performing type division on the image data of the house under construction obtained in the step S2 in a comparison clustering unsupervised mode;
it should be noted that the name of the category is not important, the number of categories of the cluster can be set at will, for example, the categories can be set as category 1, category 2, category 3, category 4, category 5, etc., and the main purpose of classifying the categories is to classify similar houses into one category.
S4, label redefinition: each building has a corresponding category after being clustered, and the category of the building data is labeled to a label on the original image;
s5, training a detection model: carrying out model training on the building data image dataset marked with the label by adopting a Yolox target detection algorithm;
s6, setting a model confidence coefficient threshold value: and adjusting the confidence threshold of each category according to the predicted effect of each category of the house data in the test set.
In this embodiment, on the test set, different confidence thresholds are set, the prediction effect on the test set is observed, then an appropriate threshold is selected for each category, and after the detection model is trained, the labels of "building in the house" are finally uniformly output, so that the algorithm can be ensured to have higher detection accuracy and recall rate, as shown in fig. 5, a manually-defined category recognition effect graph is shown, and as shown in fig. 6, the algorithm recognition effect graph is shown.
In summary, for data with complex scenes and a plurality of different states such as houses under construction, a multi-class labeling mode is adopted for data, and compared with a single class labeling mode, the multi-class labeling mode enables a model to learn data distribution more easily, so that convergence fitting is better, and the model is easier to optimize after being divided into a plurality of classes; compared with the condition of manually dividing the building into a plurality of types of labels, the condition of each period of the building under construction is abnormal, complicated and difficult to define along with the conditions of camera shooting angle, house type, house shielding and the like, the potential features among the houses are learned in an unsupervised clustering mode, and the images with high similarity are divided into one type, so that the division of a data set is more accurate, the training and convergence of a model are more facilitated, and the prediction precision of the model on the building under construction is improved; aiming at the condition that the building is divided into a plurality of categories, the prediction precision of each category in the actual scene is different, different threshold values are set for each category according to the prediction condition on the test set, and then the labels of the building are uniformly output, so that the high recall rate can be effectively maintained, and the detection precision can be effectively improved.
Example 2
As shown in fig. 3, based on embodiment 1, the performing of category classification on the building image data in step S3 in an unsupervised manner by using contrast clustering includes the following steps:
under-construction house image data sample xiUsing two data enhancement modes TaAnd TbThe data enhancement mode is randomly selected from random cutting, gray level transformation, chroma transformation, Gaussian blur and horizontal inversion, and the enhanced image
Figure RE-GDA0003589141400000081
And
Figure RE-GDA0003589141400000082
comprises the following steps:
Figure RE-GDA0003589141400000083
Figure RE-GDA0003589141400000084
and adopting Resnet34 as a feature extraction network for the enhanced image, wherein the extracted feature vector is expressed as:
Figure RE-GDA0003589141400000085
Figure RE-GDA0003589141400000086
wherein
Figure RE-GDA0003589141400000091
And
Figure RE-GDA0003589141400000092
representing a feature vector, f (-) represents a feature extraction network;
feature vector
Figure RE-GDA0003589141400000093
And
Figure RE-GDA0003589141400000094
and then mapped by two layers of nonlinear perceptrons (MLPs) as follows:
Figure RE-GDA0003589141400000095
Figure RE-GDA0003589141400000096
wherein
Figure RE-GDA0003589141400000097
And
Figure RE-GDA0003589141400000098
representing the mapped feature vector, g (-) represents the nonlinear sensor;
defining a sample
Figure RE-GDA0003589141400000099
Loss of
Figure RE-GDA00035891414000000910
Comprises the following steps:
Figure RE-GDA00035891414000000911
Figure RE-GDA00035891414000000912
wherein tau isIIs a temperature parameter of an example layer, 0.5 is taken in the experiment, s (·) represents cosine similarity operation, and N represents a set batch size, so that data enhancement is performedThere are 2N data in batch, 1 positive sample pair and 2N-2 negative sample pairs can be formed for each sample,
Figure RE-GDA00035891414000000913
and
Figure RE-GDA00035891414000000914
respectively represents that the jth sample in one batch passes through TaAnd TbExtracting feature vectors after data enhancement;
total loss of samples at the instance level linsAs shown in the following formula:
Figure RE-GDA00035891414000000915
when the data samples are mapped to the same number of spatial dimensions as the cluster, each dimension of the data features can be regarded as the probability that the sample belongs to the class, the feature vector extracted by the feature extraction network is mapped into M through two layers of nonlinear perceptors, and M represents the cluster class dimension vector, namely:
Figure RE-GDA00035891414000000916
Figure RE-GDA00035891414000000917
wherein g isc(. cndot.) is a non-linear sensor,
Figure RE-GDA00035891414000000918
and
Figure RE-GDA00035891414000000919
respectively representing the M-dimensional vectors after being mapped by the perceptron, the samples in one batch are at TaAnd TbData enhanced down-composition Ya∈RN×M,Yb∈RN×MA matrix;
definition of
Figure RE-GDA0003589141400000101
Is YaIn the ith column, the first column, similarly,
Figure RE-GDA0003589141400000102
is YbColumn i, therefore sample
Figure RE-GDA0003589141400000103
And
Figure RE-GDA0003589141400000104
loss at cluster level
Figure RE-GDA0003589141400000105
Is defined as:
Figure RE-GDA0003589141400000106
Figure RE-GDA0003589141400000107
wherein tau isCIs a parameter of temperture of cluster level, 0.9 is taken in the experiment, M represents the number of cluster types,
Figure RE-GDA0003589141400000108
for the sample at TaData enhances the cluster allocation for the jth category, and, similarly,
Figure RE-GDA0003589141400000109
indicating that the sample is at TbData-enhanced clustering assignment of the jth category;
total loss of samples at cluster level lcluIs represented by the following formula:
Figure RE-GDA00035891414000001010
h (y) assigns information entropy of probability to the clusters, so as to avoid that the network assigns all instances to one cluster, which is specifically expressed as the following formula:
Figure RE-GDA00035891414000001011
Figure RE-GDA00035891414000001012
wherein the content of the first and second substances,
Figure RE-GDA00035891414000001013
is shown at TkThe data of (1) enhances the probability that the jth sample belongs to the category i, | · | | computationally1Represents the norm L1, YkIndicates that a sample in batch is at TkAn output matrix under data enhancement;
the final model loss function is therefore l ═ lins+lclu
And through error back propagation, minimizing a loss function to obtain a clustering model, and then carrying out class division on the data set through the clustering model.

Claims (5)

1. A house under construction identification method based on unsupervised clustering is characterized by comprising the following steps:
s1, data set production: acquiring image data of a house under construction through an image acquisition device;
s2, preprocessing the data set: manually marking the position of the building under construction in the collected image data, and then cutting the marked building under construction to obtain a building image data set;
s3, unsupervised clustering of data: setting the number of the types of the clusters, and then performing type division on the image data of the house under construction obtained in the step S2 in a comparison clustering unsupervised mode;
s4, label redefinition: each building has a corresponding category after being clustered, and the category of the building data is labeled to a label on the original image;
s5, training a detection model: carrying out model training on the building data image dataset marked with the label by adopting a Yolox target detection algorithm;
s6, setting a model confidence threshold: and adjusting the confidence threshold of each category according to the predicted effect of each category of the house data in the test set.
2. The method according to claim 1, wherein in step S1, the image data of houses under construction at different times, angles and focal lengths are collected by the image collecting device.
3. The method for identifying houses under construction based on unsupervised clustering according to claim 1, wherein in step S2, the locations of houses under construction in the images are manually labeled by using a Label Image tool.
4. The method for identifying houses under construction based on unsupervised clustering according to claim 1, wherein the step S3 of performing category classification on the house under construction image data in an unsupervised manner of contrast clustering comprises the following steps:
under-construction house image data sample xiUsing two data enhancement modes TaAnd TbEnhanced images
Figure FDA0003477692160000011
And
Figure FDA0003477692160000012
comprises the following steps:
Figure FDA0003477692160000013
Figure FDA0003477692160000021
and adopting Resnet34 as a feature extraction network for the enhanced image, wherein the extracted feature vector is expressed as:
Figure FDA0003477692160000022
Figure FDA0003477692160000023
wherein
Figure FDA0003477692160000024
And
Figure FDA0003477692160000025
representing a feature vector, f (-) represents a feature extraction network;
feature vector
Figure FDA0003477692160000026
And
Figure FDA0003477692160000027
and then mapped by two layers of nonlinear perceptrons (MLPs) as follows:
Figure FDA0003477692160000028
Figure FDA0003477692160000029
wherein
Figure FDA00034776921600000210
And
Figure FDA00034776921600000211
representing the mapped feature vector, g (-) represents a nonlinear sensor;
defining a sample
Figure FDA00034776921600000212
Loss of
Figure FDA00034776921600000213
Comprises the following steps:
Figure FDA00034776921600000214
Figure FDA00034776921600000215
wherein tau isIIs a temperature parameter of an example level, s (·) represents cosine similarity operation, N represents a set size of the batch, so that after data enhancement, the total number of the batch is 2N, 1 positive sample pair and 2N-2 negative sample pairs can be formed for each sample,
Figure FDA00034776921600000216
and
Figure FDA00034776921600000217
respectively represents that the jth sample in one batch passes through TaAnd TbExtracting feature vectors after data enhancement;
total loss of samples at the instance level linsAs shown in the following formula:
Figure FDA00034776921600000218
when the data samples are mapped to the same number of spatial dimensions as the cluster, each dimension of the data features can be regarded as the probability that the sample belongs to the class, the feature vector extracted by the feature extraction network is mapped into M through two layers of nonlinear perceptors, and M represents the cluster class dimension vector, namely:
Figure FDA0003477692160000031
Figure FDA0003477692160000032
wherein g isc(. cndot.) is a non-linear sensor,
Figure FDA0003477692160000033
and
Figure FDA0003477692160000034
respectively representing the M-dimensional vectors after being mapped by the perceptron, the samples in one batch are at TaAnd TbData enhanced down-composition Ya∈RN×M,Yb∈RN×MA matrix;
definition of
Figure FDA0003477692160000035
Is YaIn the ith column, the first column, similarly,
Figure FDA0003477692160000036
is YbColumn i, therefore sample
Figure FDA0003477692160000037
And
Figure FDA0003477692160000038
loss at cluster level
Figure FDA0003477692160000039
Is defined as:
Figure FDA00034776921600000310
Figure FDA00034776921600000311
wherein tau isCIs a temperture parameter of a cluster level, M represents the number of cluster categories,
Figure FDA00034776921600000312
for the sample at TaData enhances the cluster allocation for the jth category, and, similarly,
Figure FDA00034776921600000313
indicating that the sample is at TbData-enhanced clustering assignment of the jth category;
total loss of samples at cluster level lcluIs represented by the following formula:
Figure FDA00034776921600000314
h (y) assigns probability entropy to clusters, so as to avoid the network assigning all instances to a cluster, which is specifically expressed as follows:
Figure FDA00034776921600000315
Figure FDA00034776921600000316
wherein the content of the first and second substances,
Figure FDA00034776921600000317
is shown at TkThe data of (1) enhances the probability that the jth sample belongs to the category i, | · | | computationally1Represents the norm L1, YkIndicates that a sample in batch is at TkAn output matrix under data enhancement;
the final model loss function is therefore l ═ lins+lclu
And (3) minimizing a loss function through error back propagation to obtain a clustering model, and then classifying the data set through the clustering model.
5. The method of claim 4, wherein the data enhancement mode is selected randomly from random cropping, gray scale transformation, chrominance transformation, Gaussian blur and horizontal inversion.
CN202210063160.5A 2022-01-19 2022-01-19 Method for identifying houses under construction based on unsupervised clustering Pending CN114581769A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210063160.5A CN114581769A (en) 2022-01-19 2022-01-19 Method for identifying houses under construction based on unsupervised clustering

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210063160.5A CN114581769A (en) 2022-01-19 2022-01-19 Method for identifying houses under construction based on unsupervised clustering

Publications (1)

Publication Number Publication Date
CN114581769A true CN114581769A (en) 2022-06-03

Family

ID=81770825

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210063160.5A Pending CN114581769A (en) 2022-01-19 2022-01-19 Method for identifying houses under construction based on unsupervised clustering

Country Status (1)

Country Link
CN (1) CN114581769A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114898215A (en) * 2022-06-09 2022-08-12 西南交通大学 Automatic arrangement method of sound barrier

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110232404A (en) * 2019-05-21 2019-09-13 江苏理工学院 A kind of recognition methods of industrial products surface blemish and device based on machine learning
CN110348364A (en) * 2019-07-05 2019-10-18 北京工业大学 A kind of basketball video group behavior recognition methods that Unsupervised clustering is combined with time-space domain depth network
CN112396035A (en) * 2020-12-07 2021-02-23 国网电子商务有限公司 Object detection method and device based on attention detection model
CN113312970A (en) * 2021-04-26 2021-08-27 上海工物高技术产业发展有限公司 Target object identification method, target object identification device, computer equipment and storage medium
CN113420619A (en) * 2021-06-07 2021-09-21 核工业北京地质研究院 Remote sensing image building extraction method
WO2022001489A1 (en) * 2020-06-28 2022-01-06 北京交通大学 Unsupervised domain adaptation target re-identification method
CN113903058A (en) * 2021-11-19 2022-01-07 上海玉贲智能科技有限公司 Intelligent control system based on regional personnel identification

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110232404A (en) * 2019-05-21 2019-09-13 江苏理工学院 A kind of recognition methods of industrial products surface blemish and device based on machine learning
CN110348364A (en) * 2019-07-05 2019-10-18 北京工业大学 A kind of basketball video group behavior recognition methods that Unsupervised clustering is combined with time-space domain depth network
WO2022001489A1 (en) * 2020-06-28 2022-01-06 北京交通大学 Unsupervised domain adaptation target re-identification method
CN112396035A (en) * 2020-12-07 2021-02-23 国网电子商务有限公司 Object detection method and device based on attention detection model
CN113312970A (en) * 2021-04-26 2021-08-27 上海工物高技术产业发展有限公司 Target object identification method, target object identification device, computer equipment and storage medium
CN113420619A (en) * 2021-06-07 2021-09-21 核工业北京地质研究院 Remote sensing image building extraction method
CN113903058A (en) * 2021-11-19 2022-01-07 上海玉贲智能科技有限公司 Intelligent control system based on regional personnel identification

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
YUNFAN LI等: ""Contrastive Clustering"" *
YUNFAN LI等: ""Contrastive Clustering"", 《ARXIV》, vol. 35, no. 10, pages 8549 - 8551 *
邓瑞等: ""基于深度学习的建筑"", vol. 36, no. 4, pages 20 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114898215A (en) * 2022-06-09 2022-08-12 西南交通大学 Automatic arrangement method of sound barrier

Similar Documents

Publication Publication Date Title
CN107153817B (en) Pedestrian re-identification data labeling method and device
CN111259786B (en) Pedestrian re-identification method based on synchronous enhancement of appearance and motion information of video
CN108460356B (en) Face image automatic processing system based on monitoring system
CN106096561B (en) Infrared pedestrian detection method based on image block deep learning features
CN108520226B (en) Pedestrian re-identification method based on body decomposition and significance detection
CN107133569B (en) Monitoring video multi-granularity labeling method based on generalized multi-label learning
CN111639564B (en) Video pedestrian re-identification method based on multi-attention heterogeneous network
CN110717411A (en) Pedestrian re-identification method based on deep layer feature fusion
CN111506773B (en) Video duplicate removal method based on unsupervised depth twin network
CN109919223B (en) Target detection method and device based on deep neural network
CN111611905A (en) Visible light and infrared fused target identification method
CN110728694B (en) Long-time visual target tracking method based on continuous learning
CN110728216A (en) Unsupervised pedestrian re-identification method based on pedestrian attribute adaptive learning
CN113591674B (en) Edge environment behavior recognition system for real-time video stream
CN112819065A (en) Unsupervised pedestrian sample mining method and unsupervised pedestrian sample mining system based on multi-clustering information
CN114821014A (en) Multi-mode and counterstudy-based multi-task target detection and identification method and device
CN116052212A (en) Semi-supervised cross-mode pedestrian re-recognition method based on dual self-supervised learning
CN115063832A (en) Global and local feature-based cross-modal pedestrian re-identification method for counterstudy
CN114581769A (en) Method for identifying houses under construction based on unsupervised clustering
CN111275058B (en) Safety helmet wearing and color identification method and device based on pedestrian re-identification
CN113129336A (en) End-to-end multi-vehicle tracking method, system and computer readable medium
CN107341456B (en) Weather sunny and cloudy classification method based on single outdoor color image
CN113255549B (en) Intelligent recognition method and system for behavior state of wolf-swarm hunting
CN112487927B (en) Method and system for realizing indoor scene recognition based on object associated attention
CN114445875A (en) Deep learning-based identity recognition and face comparison system and training method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination