CN114187495A - Garment fashion trend prediction method based on images - Google Patents

Garment fashion trend prediction method based on images Download PDF

Info

Publication number
CN114187495A
CN114187495A CN202210127383.3A CN202210127383A CN114187495A CN 114187495 A CN114187495 A CN 114187495A CN 202210127383 A CN202210127383 A CN 202210127383A CN 114187495 A CN114187495 A CN 114187495A
Authority
CN
China
Prior art keywords
image
clothing
foreground
fashion trend
trend prediction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210127383.3A
Other languages
Chinese (zh)
Inventor
余锋
徐硕
姜明华
周昌龙
宋坤芳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Textile University
Original Assignee
Wuhan Textile University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Textile University filed Critical Wuhan Textile University
Priority to CN202210127383.3A priority Critical patent/CN114187495A/en
Publication of CN114187495A publication Critical patent/CN114187495A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Business, Economics & Management (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Human Resources & Organizations (AREA)
  • General Engineering & Computer Science (AREA)
  • Strategic Management (AREA)
  • Economics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Game Theory and Decision Science (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Development Economics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a garment fashion trend prediction method based on images, which comprises garment image data acquisition, image foreground extraction, garment image feature extraction and garment fashion trend prediction. Firstly, collecting a clothing image data set, and preprocessing clothing image data; then obtaining a foreground image, and extracting clothing features based on a multi-convolution kernel deep neural network; and finally, a garment fashion trend prediction method based on deep learning is adopted, and the garment image characteristics are used as the input of the model to obtain the current garment fashion trend. The method can greatly reduce the calculation cost and the system complexity, promote the intellectualization of fashion trend prediction in the fashion field, and improve the effect and the quality of fashion prediction.

Description

Garment fashion trend prediction method based on images
Technical Field
The invention belongs to the technical field of intelligent clothes, and particularly relates to a clothes fashion trend prediction method based on images.
Background
At present, in the field of online clothing, designers usually design new clothes through learning experience, each time a piece of clothes is designed, a great deal of time and energy are consumed, the designers cannot possibly design clothes of each required style, popular clothes in various areas in the future cannot be easily predicted, and a plurality of designers familiar with the areas are usually required to participate. Therefore, in the field of clothing, intelligent prediction of future development trends of clothing has a potential and huge application scene.
Chinese patent publication No. CN110705755A discloses "a garment popularity trend prediction method and apparatus based on deep learning", which collects popular garment pictures and information from e-commerce website over the year, performs feature extraction and integration, and outputs a garment ranking scheme with garment popularity topk according to model results, but the scheme is not accurate for garment popularity trend prediction and needs further optimization.
Disclosure of Invention
In view of the above defects or improvement needs of the prior art, the present invention provides an image-based garment popularity trend prediction method, which aims to predict garment popularity trend by a deep learning method by collecting garment pictures of current garment shopping websites on the internet, and is reliable and real-time.
To achieve the above object, according to one aspect of the present invention, there is provided an image-based clothing fashion trend prediction method, including the steps of:
step 1, firstly, a clothing image data set is collected, and the clothing image data is preprocessed;
step2, extracting the clothing image foreground by using a foreground extraction model based on image multi-scale decomposition;
step 3, extracting and fusing clothing image features based on the multi-convolution kernel deep convolution neural network to obtain a final clothing image feature map;
step4, constructing a clothing fashion trend prediction model, and taking the final clothing image characteristics as the input of the clothing fashion trend prediction model to obtain the current clothing fashion trend;
the clothing fashion trend prediction module comprises: an adaptive weighted pooling layer, a full link layer, and a softmax layer.
Further, in step 1, clothing images of all shopping websites are collected through a web crawler and a manual collection mode, wherein the shopping websites comprise amazon online shopping malls, tianmao malls, tabby and jingdong malls.
Further, the garment image data preprocessing comprises: and adjusting the size of the image by a bilinear interpolation method, and then carrying out image scale normalization and image standardization.
Further, the specific implementation manner of step2 is as follows;
step21, carrying out multi-scale decomposition on the image by using total variation to obtain a series of smooth images;
step22, representing the foreground color distribution of the given smooth image as a Gaussian mixture model, and optimizing the number of Gaussian functions of the Gaussian mixture model by using a histogram shape analysis method;
and Step23, designing an iteration termination condition according to the segmentation results of different smooth images, so that the foreground is extracted from the decomposition scale of the smooth images.
Further, the specific implementation manner of optimizing the gaussian mixture model by using a histogram shape analysis method in Step22 is as follows;
the gaussian distribution of the mth zone is expressed as follows,
Figure 471571DEST_PATH_IMAGE001
in the formula, G represents a Gaussian function,μ m sum ΣmRespectively the mean vector and covariance matrix of the color distribution of the region,u(i) Representing smooth imagesuTo (1)iPixel value is taken when calculating; det is a mathematical function used for solving a determinant of a square matrix;
using each peak of histogram to represent brightness distribution of image region, smoothing with median filter to obtain a smoothed histogram
Figure 200624DEST_PATH_IMAGE002
The histogram has 256 values, respectively
Figure 249351DEST_PATH_IMAGE003
The method comprises the steps of dividing an image into N regions by utilizing the troughs of a histogram, and calculating the number of the regions and pixels in a foreground F and a background B by combining a segmentation curveu(i) The probability of belonging to the foreground or background is:
Figure 463688DEST_PATH_IMAGE004
in the formula,
Figure 92116DEST_PATH_IMAGE005
the result of the segmentation is represented, wherein,x n =1 represents the foreground of the image,x n =0 is for the background, and the background,L F andL B respectively represent pixels areu(i) The likelihood of the foreground and the background,ω F andω B parameters representing the foreground and the background respectively,n F which represents the number of foreground pixels,n B representing the number of background pixels; the optimized gaussian mixture model is expressed as:
Figure 206834DEST_PATH_IMAGE006
wherein,U(x,w,u) As a parameterwLower partxFor smooth imageuEvaluating the segmentation result of (1);wrepresenting the front and background color distribution parameters.
Further, combining CrabCut and a color distribution model of a smooth image in Step23, converting foreground extraction into joint optimization of segmentation and decomposition scales, and extracting energy functional of foregroundx*Comprises the following steps:
Figure 832987DEST_PATH_IMAGE007
in the formula,αandβrepresenting a weight; first itemM(u,u 0 ) Is a multi-scale decomposition of the image,u 0 which represents the original image or images of the original image,uis a smooth image; the second term is the foreground extraction of the smoothed image,S(x,w,u) Expressed as:
Figure 531690DEST_PATH_IMAGE008
in the formula,U(x,w,u) As a parameterwLower partxFor smooth imageuEvaluating the segmentation result of (1);wrepresenting the color distribution parameters of the front and the background,
Figure 963809DEST_PATH_IMAGE005
the result of the segmentation is represented, wherein,x n =1 represents the foreground of the image,x n =0 represents background;V(x, u) Defined as the penalty of placing the segmentation curve on the foreground boundary as follows:
Figure 667454DEST_PATH_IMAGE009
in the formula,A i is the firstiA set of adjacent pixels of the one pixel,jis composed ofA i A pixel of (1);dis(∙) representing the Euclidean distance of the pixel pairs; [ ∙]Is an indicator function;γandβrepresenting a weight; u (∙) represents a smoothed image.
Further, the multi-convolution kernel deep convolution neural network comprises a multi-convolution kernel feature extraction module and a multi-convolution kernel feature fusion module, and the specific implementation mode is as follows;
(31) extracting clothing image features including style, color system and style by using a multi-convolution kernel feature extraction module;
firstly, performing convolution operation and activation function operation twice on an input image, extracting image features to generate a clothing image feature map with the dimension of 224 multiplied by 64, performing maximum pooling operation on the extracted feature image on the basis, and converting the feature map dimension to 112 multiplied by 64; then, performing convolution operation and activation function operation, performing maximum pooling operation on the feature maps extracted after corresponding operation, and generating a feature map with dimensions of 56 × 56 × 128 as the input of the multi-convolution kernel fusion module;
(32) fusing the extracted features by using a multi-convolution kernel feature fusion module to obtain a final clothing image feature map
The multi-convolution kernel feature fusion module comprises: intra-module feature information fusion and inter-module feature information fusion;
wherein, the intra-module feature information fusion comprises three branches, the three branches respectively use convolution kernel with the sizes of 3 multiplied by 3, 5 multiplied by 5 and 7 multiplied by 7 to further extract the features output by the multi-convolution kernel feature extraction module, the three branches use a parallel mode to extract the features, the inter-module feature information fusion respectively carries out 3 multiplied by 3 convolution operation after aggregating the three branches in the intra-module feature information fusion, then the three branches are input into the intra-module feature information fusion part to aggregate the clothing feature information again, finally the extracted feature information is fused and aggregated with the input feature information again through 1 multiplied by 1 convolution operation, thereby achieving the purpose of fusing the extracted feature information, in the above convolution operations, all convolution operations except the 1 × 1 convolution operation are followed by the Relu activation function.
Further, the specific processing procedure of the adaptive weighting pooling layer comprises;
inputting: feature map to be pooled, pooling window sizenLoss functionJLearning rateβ
Step 41: for each pooling layer, selecting the number of importance parameters according to the size of the pooling window of the layer, whereinnCharacteristic valueα i Pooling window random initializationnAn importance parameterk i i=1,2,...,n
Step 42: sorting the characteristic values in each pooling window from large to small to obtain
Figure 526825DEST_PATH_IMAGE010
Step 43: performing softmax normalization on the initialized importance parameters to obtain weight parameters;
Figure 467753DEST_PATH_IMAGE011
step 44: multiplying the weight parameter by the corresponding characteristic value in each pooling window, and accumulating to obtain a pooling result:
Figure 703563DEST_PATH_IMAGE012
step 45: initialized weight parameterw i And continuously iterating and optimizing through gradient reduction along with the progress of back propagation in the training process until convergence:
Figure 527293DEST_PATH_IMAGE013
wherein,α i in order to be a characteristic value of the image,k i the parameters are randomly initialized for the pooling window,w i is a weight parameter, z is a weight parameter multiplied by the corresponding characteristic value in each pooling window and then accumulated to obtain a pooling result,representing partial differentiation.
Further, the loss function of the whole garment fashion trend prediction model and the loss function in the self-adaptive weighting pooling layerJLikewise, a cross entropy loss function is employed:
Figure 557566DEST_PATH_IMAGE014
wherein,xclothing image features representing input modelThe figure is a figure of merit,pandqrespectively representing the classification real value and the clothing classification predicted value of the clothing classification.
Further, the processing procedure of the softmax layer is as follows;
Figure 965283DEST_PATH_IMAGE015
wherein,Z i is as followsiThe output value of each node, C, is the number of output nodes, i.e. the number of categories of the final classification result.
According to another aspect of the invention, an image-based garment fashion trend prediction system is provided, which comprises the following modules:
the clothing image data acquisition module is used for collecting a clothing image data set and preprocessing clothing image data;
the image foreground extraction module is used for extracting the clothing image foreground by using a foreground extraction model based on image multi-scale decomposition;
the clothing image feature extraction module is used for extracting and fusing clothing image features based on the multi-convolution kernel deep convolution neural network to obtain a final clothing image feature map;
the garment fashion trend prediction module is used for constructing a garment fashion trend prediction model, and the final garment image characteristics are used as the input of the garment fashion trend prediction model to obtain the current garment fashion trend;
the garment fashion trend prediction model comprises: an adaptive weighted pooling layer, a full link layer, and a softmax layer.
In general, compared with the prior art, the above technical solution contemplated by the present invention can achieve the following beneficial effects:
(1) the garment fashion trend prediction method based on the images, provided by the invention, is used for predicting the garment fashion trend by utilizing the garment images, wherein the garment fashion trend comprises garment style, color system and style, and the garment fashion trend is predicted by a deep learning method, so that the garment fashion trend prediction method is reliable and has real-time performance;
(2) compared with the prior art, the garment popularity trend prediction method based on the images can greatly reduce the calculation cost, reduce the system complexity and improve the popularity prediction effect and quality.
Drawings
FIG. 1 is a schematic flow chart of a system for predicting fashion trends of clothing based on images according to an embodiment of the present invention;
fig. 2 is a structural diagram of a deep convolutional neural network with multiple convolutional kernels according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. In addition, the technical features involved in the embodiments of the present invention described below may be combined with each other as long as they do not conflict with each other.
Fig. 1 is a schematic flow diagram of a garment fashion trend prediction system based on an image according to an embodiment, and the system includes 4 parts, namely a garment image data acquisition module, an image foreground extraction module, a garment image feature extraction module, and a garment fashion trend prediction module, and the specific processing procedures of the modules are as follows:
the clothing image data acquisition module is used for collecting a clothing image data set and preprocessing clothing image data;
the image foreground extraction module is used for extracting the clothing image foreground by using a foreground extraction model based on image multi-scale decomposition;
the clothing image feature extraction module is used for extracting and fusing clothing image features based on the multi-convolution kernel deep convolution neural network to obtain a final clothing image feature map;
the garment fashion trend prediction module is used for constructing a garment fashion trend prediction model, and the final garment image characteristics are used as the input of the garment fashion trend prediction model to obtain the current garment fashion trend;
the garment fashion trend prediction model comprises: an adaptive weighted pooling layer, a full link layer, and a softmax layer.
Corresponding to the system, the invention also provides a garment fashion trend prediction method based on the image, which comprises the following steps:
(1) firstly, collecting a clothing image data set, and preprocessing clothing image data;
in this embodiment, the clothing images of the large shopping websites are collected by a web crawler and a manual collection method, and the clothing images are subjected to image scale normalization and image standardization, wherein the shopping websites include amazon-online shopping mall (amazon.com), makitar mall (tmall.com), Taobao.com and Jingdong mall (jd.com).
Wherein, clothing image data preprocessing includes: and adjusting the size of the image by a bilinear interpolation method, and then carrying out image scale normalization and image standardization. In an embodiment, the image size is 224 × 224 × 3.
(2) Extracting an image foreground for obtaining a foreground image;
and extracting the clothing image foreground by using a foreground extraction model based on image multi-scale decomposition.
Step 21: carrying out multi-scale decomposition on the image by using total variation to obtain a series of smooth images, wherein the decomposition protects the edges of the image, smoothes the texture and compresses the distribution range of the colors of the image region;
step 22: expressing the foreground color distribution of the given smooth image as a Gaussian mixture model, and optimizing the number of Gaussian functions of the Gaussian mixture model by using a histogram shape analysis method;
and for each smooth image, accurately modeling the color distribution of the image by adopting a histogram shape analysis method.
Wherein the histogram shape analysis method optimizes a gaussian mixture model. Assuming that the color distribution of each region in the smoothed image is compact, the color distribution of the region can be expressed as a gaussian function, taking the mth region as an example:
Figure 739204DEST_PATH_IMAGE016
in the formula, G represents a Gaussian function,μ m sum ΣmRespectively the mean vector and covariance matrix of the color distribution of the region,u(i) Representing smooth imagesuTo (1)iPixel value is taken when calculating; det is a mathematical function used for solving a determinant of a square matrix;
using each peak of histogram to represent brightness distribution of image region, smoothing with median filter to obtain a smoothed histogram
Figure 683020DEST_PATH_IMAGE002
The histogram has 256 values, respectively
Figure 884194DEST_PATH_IMAGE003
The method comprises the steps of dividing an image into N regions by utilizing the troughs of a histogram, and calculating the number of the regions and pixels in a foreground F and a background B by combining a segmentation curveu(i) The probability of belonging to the foreground or background is:
Figure 782136DEST_PATH_IMAGE017
in the formula,
Figure 297431DEST_PATH_IMAGE005
the result of the segmentation is represented, wherein,x n =1 represents the foreground of the image,x n =0 is for the background, and the background,L F andL B respectively represent pixels areu(i) The likelihood of the foreground and the background,ω F andω B parameters representing the foreground and the background respectively,n F which represents the number of foreground pixels,n B representing the number of background pixels; the optimized gaussian mixture model is expressed as:
Figure 345022DEST_PATH_IMAGE018
wherein,U(x,w,u) As a parameterwLower partxFor smooth imageuEvaluating the segmentation result of (1);wrepresenting the front and background color distribution parameters.
Step 23: and designing an iteration termination condition according to the segmentation results of different smooth images so as to extract the foreground from the decomposition scale of the smooth image.
And combining CrabCut and a color distribution model of the smooth image, and converting foreground extraction into joint optimization of segmentation and decomposition scale. An original image with N pixels
Figure 467830DEST_PATH_IMAGE019
Divided by an initial rectangular frame into a background region B and a foreground region F with a small number of background pixels, RGB representing red green blue respectively,u R which represents a set of red pixels, is shown,u G a set of green pixels is represented as,u B representing a set of blue pixels that are to be addressed,
Figure 866450DEST_PATH_IMAGE020
representing an original imageu 0 The set of all pixels in. Energy functional with its foreground extractedx*Comprises the following steps:
Figure 231441DEST_PATH_IMAGE007
in the formula,αandβrepresenting a weight; first itemM(u,u 0 ) Is a multi-scale decomposition of the image,u 0 which represents the original image or images of the original image,uis a smooth image; the second term is the foreground extraction of the smoothed image,S(x,w,u) Expressed as:
Figure 336801DEST_PATH_IMAGE021
in the formula,U(x,w,u) As a parameterwLower partxFor smooth imageuEvaluating the segmentation result of (1);wrepresenting the color distribution parameters of the front and the background,
Figure 630510DEST_PATH_IMAGE005
the result of the segmentation is represented, wherein,x n =1 represents the foreground of the image,x n =0 represents background;V(x, u) Defined as the penalty of placing the segmentation curve on the foreground boundary as follows:
Figure 985268DEST_PATH_IMAGE022
in the formula,A i is the firstiA set of adjacent pixels of the one pixel,jis composed ofA i A pixel of (1);dis(∙) representing the Euclidean distance of the pixel pairs; [ ∙]Is an indicator function;γandβrepresenting a weight; u (∙) represents a smoothed image.
In a particular embodiment of the present invention,γtaking 50, to ensure that the energy of the above formula is greater at low gradients and less at high gradients,
Figure 151020DEST_PATH_IMAGE023
where 〈 ∙ 〉 represents the mean value.
(3) Clothing image feature extraction and fusion are carried out on the basis of a multi-convolution kernel deep convolution neural network to obtain a final clothing image feature map, wherein the multi-convolution kernel deep convolution neural network comprises a multi-convolution kernel feature extraction module and a multi-convolution kernel feature fusion module;
(31) and extracting the clothing image features including style, color system and style by using a multi-convolution kernel feature extraction module.
In a specific embodiment, firstly, the convolution operation and the activation function operation are performed twice on the input image, the image features are extracted, a clothing image feature map with the dimension of 224 × 224 × 64 is generated, and on the basis, the maximum pooling operation is performed on the extracted feature image, and the feature map dimension is converted into 112 × 112 × 64. Then, performing convolution operation and activation function operation, and performing maximum pooling operation on the feature maps extracted after corresponding operation to generate feature maps with dimensions of 56 × 56 × 128 as input of the multi-convolution kernel fusion module.
(32) And fusing the extracted features by using a multi-convolution kernel feature fusion module to obtain a final clothing image feature map.
As shown in fig. 2, it is a network structure diagram of a multi-convolution kernel feature fusion module provided in the embodiment, where the multi-convolution kernel feature fusion module includes: intra-module feature information fusion and inter-module feature information fusion.
Wherein, the intra-module feature information fusion comprises three branches, the three branches respectively use convolution kernel with the sizes of 3 multiplied by 3, 5 multiplied by 5 and 7 multiplied by 7 to further extract the features output by the multi-convolution kernel feature extraction module, the three branches use a parallel mode to extract the features, the inter-module feature information fusion respectively carries out 3 multiplied by 3 convolution operation after aggregating the three branches in the intra-module feature information fusion, then the three branches are input into the intra-module feature information fusion part to aggregate the clothing feature information again, finally the extracted feature information is fused and aggregated with the input feature information again through 1 multiplied by 1 convolution operation, thereby achieving the purpose of fusing the extracted feature information, in the above convolution operations, all convolution operations except the 1 × 1 convolution operation are followed by the Relu activation function.
(4) And constructing a garment fashion trend prediction model, and taking the final garment image characteristics as the input of the garment fashion trend prediction model to obtain the current garment fashion trend.
Wherein the clothing fashion trend prediction module comprises: an adaptive weighted pooling layer, a full link layer, and softmax.
The processing procedure of the self-adaptive weighting pooling layer specifically comprises the following steps:
inputting: feature map to be pooled, pooling window sizenLoss functionJLearning rateβ
Step 41: for each pooling layer, selecting the number of importance parameters according to the size of the pooling window of the layer, whereinnCharacteristic valueα i Pooling window random initializationnAn importance parameterk i i=1,2,...,n
Step42:Sorting the characteristic values in each pooling window from large to small to obtain
Figure 173203DEST_PATH_IMAGE010
Step 43: performing softmax normalization on the initialized importance parameters to obtain weight parameters;
Figure 637813DEST_PATH_IMAGE024
step 44: multiplying the weight parameter by the corresponding characteristic value in each pooling window, and accumulating to obtain a pooling result:
Figure 745447DEST_PATH_IMAGE025
step 45: initialized weight parameterw i And continuously iterating and optimizing through gradient reduction along with the progress of back propagation in the training process until convergence:
Figure 452240DEST_PATH_IMAGE013
wherein,α i in order to be a characteristic value of the image,k i the parameters are randomly initialized for the pooling window,w i is a weight parameter, z is a weight parameter multiplied by the corresponding characteristic value in each pooling window and then accumulated to obtain a pooling result,representing partial differentiation.
Wherein, the loss function of the whole garment fashion trend prediction model and the loss function in the self-adaptive weighting pooling layerJLikewise, a cross entropy loss function is employed:
Figure 63350DEST_PATH_IMAGE026
wherein,xa clothing image feature map representing the input model,pandqclassification true value and clothing classification respectively representing clothing classificationAnd (4) predicting the class. The numerical value obtained by calculating the cross entropy loss function does not necessarily meet the condition and significance of probability distribution, so that the data is processed into a probability distribution form by finally activating the function by softmax, and the requirement of a multi-classification task of a clothing image algorithm is met.
Figure 964441DEST_PATH_IMAGE015
Wherein,Z i is as followsiThe output value of each node, C, is the number of output nodes, i.e. the number of categories of the final classification result.
It will be understood by those skilled in the art that the foregoing is only a preferred embodiment of the present invention, and is not intended to limit the invention, and that any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the scope of the present invention.

Claims (10)

1. An image-based garment fashion trend prediction method is characterized by comprising the following steps:
step 1, firstly, a clothing image data set is collected, and the clothing image data is preprocessed;
step2, extracting the clothing image foreground by using a foreground extraction model based on image multi-scale decomposition;
step 3, extracting and fusing clothing image features based on the multi-convolution kernel deep convolution neural network to obtain a final clothing image feature map;
step4, constructing a clothing fashion trend prediction model, and taking the final clothing image characteristics as the input of the clothing fashion trend prediction model to obtain the current clothing fashion trend;
the clothing fashion trend prediction module comprises: an adaptive weighted pooling layer, a full link layer, and a softmax layer.
2. The image-based garment fashion trend prediction method of claim 1, characterized in that: in the step 1, clothing images of all large shopping websites are collected through a web crawler and a manual collection mode, wherein the shopping websites comprise an Amazon-online shopping mall, a Temple mall, a Taobao net and a Jingdong mall;
the garment image data preprocessing comprises the following steps: and adjusting the size of the image by a bilinear interpolation method, and then carrying out image scale normalization and image standardization.
3. The image-based garment fashion trend prediction method of claim 1, characterized in that: the specific implementation manner of the step2 is as follows;
step21, carrying out multi-scale decomposition on the image by using total variation to obtain a series of smooth images;
step22, representing the foreground color distribution of the given smooth image as a Gaussian mixture model, and optimizing the number of Gaussian functions of the Gaussian mixture model by using a histogram shape analysis method;
and Step23, designing an iteration termination condition according to the segmentation results of different smooth images, so that the foreground is extracted from the decomposition scale of the smooth images.
4. The image-based garment fashion trend prediction method of claim 3, characterized in that: the concrete implementation mode of optimizing the Gaussian mixture model by using a histogram shape analysis method in Step22 is as follows;
the gaussian distribution of the mth zone is expressed as follows,
Figure 441246DEST_PATH_IMAGE001
in the formula, G represents a Gaussian function,μ m sum ΣmRespectively the mean vector and covariance matrix of the color distribution of the region,u(i) Representing smooth imagesuTo (1)iPixel value is taken when calculating; det is a mathematical function used for solving a determinant of a square matrix;
using each peak of histogram to represent brightness distribution of image region, and smoothing by median filteringThe smoothed histogram is
Figure 278752DEST_PATH_IMAGE002
The histogram has 256 values, respectively
Figure 31813DEST_PATH_IMAGE003
The method comprises the steps of dividing an image into N regions by utilizing the troughs of a histogram, and calculating the number of the regions and pixels in a foreground F and a background B by combining a segmentation curveu(i) The probability of belonging to the foreground or background is:
Figure 271165DEST_PATH_IMAGE004
in the formula,
Figure 815760DEST_PATH_IMAGE005
the result of the segmentation is represented, wherein,x n =1 represents the foreground of the image,x n =0 is for the background, and the background,L F andL B respectively represent pixels areu(i) The likelihood of the foreground and the background,ω F andω B parameters representing the foreground and the background respectively,n F which represents the number of foreground pixels,n B representing the number of background pixels; the optimized gaussian mixture model is expressed as:
Figure 874983DEST_PATH_IMAGE006
wherein,U(x,w,u) As a parameterwLower partxFor smooth imageuEvaluating the segmentation result of (1);wrepresenting the front and background color distribution parameters.
5. The image-based garment fashion trend prediction method of claim 4, characterized in that: color distribution model combining CrabCut and smooth image in Step23Type, converting foreground extraction into joint optimization of segmentation and decomposition scale, energy functional of foreground extractionx*Comprises the following steps:
Figure 979205DEST_PATH_IMAGE007
in the formula,αandβrepresenting a weight; first itemM(u,u 0 ) Is a multi-scale decomposition of the image,u 0 which represents the original image or images of the original image,uis a smooth image; the second term is the foreground extraction of the smoothed image,S(x,w,u) Expressed as:
Figure 260014DEST_PATH_IMAGE008
in the formula,U(x,w,u) As a parameterwLower partxFor smooth imageuEvaluating the segmentation result of (1);wrepresenting the color distribution parameters of the front and the background,
Figure 791489DEST_PATH_IMAGE005
the result of the segmentation is represented, wherein,x n =1 represents the foreground of the image,x n =0 represents background;V(x,u) Defined as the penalty of placing the segmentation curve on the foreground boundary as follows:
Figure 587276DEST_PATH_IMAGE009
in the formula,A i is the firstiA set of adjacent pixels of the one pixel,jis composed ofA i A pixel of (1);dis(∙) representing the Euclidean distance of the pixel pairs; [ ∙]Is an indicator function;γandβrepresenting a weight; u (∙) represents a smoothed image.
6. The image-based garment fashion trend prediction method of claim 1, characterized in that: the multi-convolution kernel deep convolution neural network comprises a multi-convolution kernel feature extraction module and a multi-convolution kernel feature fusion module, and the specific implementation mode is as follows;
(31) extracting clothing image features including style, color system and style by using a multi-convolution kernel feature extraction module;
firstly, performing convolution operation and activation function operation twice on an input image, extracting image features to generate a clothing image feature map with the dimension of 224 multiplied by 64, performing maximum pooling operation on the extracted feature image on the basis, and converting the feature map dimension to 112 multiplied by 64; then, performing convolution operation and activation function operation, performing maximum pooling operation on the feature maps extracted after corresponding operation, and generating a feature map with dimensions of 56 × 56 × 128 as the input of the multi-convolution kernel fusion module;
(32) fusing the extracted features by using a multi-convolution kernel feature fusion module to obtain a final clothing image feature map
The multi-convolution kernel feature fusion module comprises: intra-module feature information fusion and inter-module feature information fusion;
wherein, the intra-module feature information fusion comprises three branches, the three branches respectively use convolution kernel with the sizes of 3 multiplied by 3, 5 multiplied by 5 and 7 multiplied by 7 to further extract the features output by the multi-convolution kernel feature extraction module, the three branches use a parallel mode to extract the features, the inter-module feature information fusion respectively carries out 3 multiplied by 3 convolution operation after aggregating the three branches in the intra-module feature information fusion, then the three branches are input into the intra-module feature information fusion part to aggregate the clothing feature information again, finally the extracted feature information is fused and aggregated with the input feature information again through 1 multiplied by 1 convolution operation, thereby achieving the purpose of fusing the extracted feature information, in the above convolution operations, all convolution operations except the 1 × 1 convolution operation are followed by the Relu activation function.
7. The image-based garment fashion trend prediction method of claim 1, characterized in that: the specific processing procedure of the self-adaptive weighting pooling layer comprises the following steps of;
inputting: feature map to be pooled, pooling window sizenLoss functionJLearning rateβ
Step 41: for each pooling layer, selecting the number of importance parameters according to the size of the pooling window of the layer, whereinnCharacteristic valueα i Pooling window random initializationnAn importance parameterk i i=1,2,...,n
Step 42: sorting the characteristic values in each pooling window from large to small to obtain
Figure 432872DEST_PATH_IMAGE010
Step 43: performing softmax normalization on the initialized importance parameters to obtain weight parameters;
Figure 630504DEST_PATH_IMAGE011
step 44: multiplying the weight parameter by the corresponding characteristic value in each pooling window, and accumulating to obtain a pooling result:
Figure 536143DEST_PATH_IMAGE012
step 45: initialized weight parameterw i And continuously iterating and optimizing through gradient reduction along with the progress of back propagation in the training process until convergence:
Figure 618893DEST_PATH_IMAGE013
wherein,α i in order to be a characteristic value of the image,k i the parameters are randomly initialized for the pooling window,w i is a weight parameter, and z is a weight parameter corresponding to the feature in each pooling windowThe values are multiplied and then accumulated to obtain a pooling result,representing partial differentiation.
8. The image-based garment fashion trend prediction method of claim 7, characterized in that: loss function of whole garment fashion trend prediction model and loss function in adaptive weighted pooling layerJLikewise, a cross entropy loss function is employed:
Figure 268180DEST_PATH_IMAGE014
wherein,xa clothing image feature map representing the input model,pandqrespectively representing the classification real value and the clothing classification predicted value of the clothing classification.
9. The image-based garment fashion trend prediction method of claim 1, characterized in that: the processing procedure of the softmax layer is as follows;
Figure 54739DEST_PATH_IMAGE015
wherein,Z i is as followsiThe output value of each node, C, is the number of output nodes, i.e. the number of categories of the final classification result.
10. An image-based garment fashion trend prediction system is characterized by comprising the following modules:
the clothing image data acquisition module is used for collecting a clothing image data set and preprocessing clothing image data;
the image foreground extraction module is used for extracting the clothing image foreground by using a foreground extraction model based on image multi-scale decomposition;
the clothing image feature extraction module is used for extracting and fusing clothing image features based on the multi-convolution kernel deep convolution neural network to obtain a final clothing image feature map;
the garment fashion trend prediction module is used for constructing a garment fashion trend prediction model, and the final garment image characteristics are used as the input of the garment fashion trend prediction model to obtain the current garment fashion trend;
the garment fashion trend prediction model comprises: an adaptive weighted pooling layer, a full link layer, and a softmax layer.
CN202210127383.3A 2022-02-11 2022-02-11 Garment fashion trend prediction method based on images Pending CN114187495A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210127383.3A CN114187495A (en) 2022-02-11 2022-02-11 Garment fashion trend prediction method based on images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210127383.3A CN114187495A (en) 2022-02-11 2022-02-11 Garment fashion trend prediction method based on images

Publications (1)

Publication Number Publication Date
CN114187495A true CN114187495A (en) 2022-03-15

Family

ID=80545831

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210127383.3A Pending CN114187495A (en) 2022-02-11 2022-02-11 Garment fashion trend prediction method based on images

Country Status (1)

Country Link
CN (1) CN114187495A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116188792A (en) * 2023-02-23 2023-05-30 四川大学 Quantitative analysis method and system for whole blood cell scatter diagram

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103718166A (en) * 2011-08-02 2014-04-09 索尼公司 Information processing apparatus, information processing method, and computer program product
CN108960499A (en) * 2018-06-27 2018-12-07 东华大学 A kind of Fashion trend predicting system merging vision and non-vision feature
CN110705755A (en) * 2019-09-07 2020-01-17 创新奇智(广州)科技有限公司 Garment fashion trend prediction method and device based on deep learning
CN112434210A (en) * 2020-12-14 2021-03-02 武汉纺织大学 Clothing fashion trend prediction system and method
CN112819510A (en) * 2021-01-21 2021-05-18 江阴逐日信息科技有限公司 Fashion trend prediction method, system and equipment based on clothing multi-attribute recognition
CN113159826A (en) * 2020-12-28 2021-07-23 武汉纺织大学 Garment fashion element prediction system and method based on deep learning

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103718166A (en) * 2011-08-02 2014-04-09 索尼公司 Information processing apparatus, information processing method, and computer program product
CN108960499A (en) * 2018-06-27 2018-12-07 东华大学 A kind of Fashion trend predicting system merging vision and non-vision feature
CN110705755A (en) * 2019-09-07 2020-01-17 创新奇智(广州)科技有限公司 Garment fashion trend prediction method and device based on deep learning
CN112434210A (en) * 2020-12-14 2021-03-02 武汉纺织大学 Clothing fashion trend prediction system and method
CN113159826A (en) * 2020-12-28 2021-07-23 武汉纺织大学 Garment fashion element prediction system and method based on deep learning
CN112819510A (en) * 2021-01-21 2021-05-18 江阴逐日信息科技有限公司 Fashion trend prediction method, system and equipment based on clothing multi-attribute recognition

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
王斌 等: ""基于图像多尺度分解的前景提取"", 《四川大学学报(自然科学版)》 *
赵长乐: ""基于卷积神经网络的服装图像分类与去噪研究"", 《万方数据》 *
龚柯: ""基于卷积神经网络的服装分类算法研究"", 《中国优秀硕士学位论文全文数据库 (工程科技Ⅰ辑)》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116188792A (en) * 2023-02-23 2023-05-30 四川大学 Quantitative analysis method and system for whole blood cell scatter diagram
CN116188792B (en) * 2023-02-23 2023-10-20 四川大学 Quantitative analysis method and system for whole blood cell scatter diagram

Similar Documents

Publication Publication Date Title
CN111476713B (en) Intelligent weather image identification method and system based on multi-depth convolution neural network fusion
CN104182772B (en) A kind of gesture identification method based on deep learning
CN108573276A (en) A kind of change detecting method based on high-resolution remote sensing image
CN111899172A (en) Vehicle target detection method oriented to remote sensing application scene
CN107133943A (en) A kind of visible detection method of stockbridge damper defects detection
CN108510012A (en) A kind of target rapid detection method based on Analysis On Multi-scale Features figure
CN112396160A (en) Transaction fraud detection method and system based on graph neural network
CN109558806A (en) The detection method and system of high score Remote Sensing Imagery Change
CN110728179A (en) Pig face identification method adopting multi-path convolutional neural network
CN115249332B (en) Hyperspectral image classification method and device based on space spectrum double-branch convolution network
CN111178121B (en) Pest image positioning and identifying method based on spatial feature and depth feature enhancement technology
CN103390279A (en) Target prospect collaborative segmentation method combining significant detection and discriminant study
CN111639719A (en) Footprint image retrieval method based on space-time motion and feature fusion
CN107506792B (en) Semi-supervised salient object detection method
CN104063686A (en) System and method for performing interactive diagnosis on crop leaf segment disease images
CN113420794B (en) Binaryzation Faster R-CNN citrus disease and pest identification method based on deep learning
CN114863173B (en) Self-mutual-attention hyperspectral image classification method for land resource audit
CN114663769B (en) Fruit identification method based on YOLO v5
CN117876890B (en) Multi-source remote sensing image classification method based on multi-level feature fusion
Li et al. SGML: A symmetric graph metric learning framework for efficient hyperspectral image classification
CN114187495A (en) Garment fashion trend prediction method based on images
CN112967296B (en) Point cloud dynamic region graph convolution method, classification method and segmentation method
CN104933410B (en) A kind of high spectrum image spectral domain and spatial domain joint classification method
CN117853596A (en) Unmanned aerial vehicle remote sensing mapping method and system
CN108765384B (en) Significance detection method for joint manifold sequencing and improved convex hull

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20220315

RJ01 Rejection of invention patent application after publication