CN109815873A - Merchandise display method, apparatus, equipment and medium based on image recognition - Google Patents

Merchandise display method, apparatus, equipment and medium based on image recognition Download PDF

Info

Publication number
CN109815873A
CN109815873A CN201910042541.3A CN201910042541A CN109815873A CN 109815873 A CN109815873 A CN 109815873A CN 201910042541 A CN201910042541 A CN 201910042541A CN 109815873 A CN109815873 A CN 109815873A
Authority
CN
China
Prior art keywords
image
recognized
images
commodity
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910042541.3A
Other languages
Chinese (zh)
Inventor
罗琳耀
徐国强
邱寒
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
OneConnect Smart Technology Co Ltd
Original Assignee
OneConnect Smart Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by OneConnect Smart Technology Co Ltd filed Critical OneConnect Smart Technology Co Ltd
Priority to CN201910042541.3A priority Critical patent/CN109815873A/en
Publication of CN109815873A publication Critical patent/CN109815873A/en
Priority to PCT/CN2019/120988 priority patent/WO2020147430A1/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions

Landscapes

  • Business, Economics & Management (AREA)
  • Accounting & Taxation (AREA)
  • Finance (AREA)
  • Development Economics (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Strategic Management (AREA)
  • Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The present invention discloses a kind of merchandise display method, apparatus, equipment and medium based on image recognition, this method includes carrying out recognition of face and cluster at least two frame images to be recognized in the target video data of commodity to be presented using Face datection model, obtains the corresponding customer quantity of target video data and the corresponding image clustering set of each client;If customer quantity is greater than preset quantity, the images to be recognized in each image clustering set is identified using preparatory trained micro- Expression Recognition model, obtains the single frames mood of each images to be recognized;Single frames mood based on images to be recognized obtains the target emotion of client corresponding with image clustering set;According to the target emotion of customer quantity and the corresponding client of each image clustering set, final mood is obtained;According to the corresponding final mood of commodity to be presented, intended display commodity are obtained, to solve the problems, such as to show that commodity attraction is inadequate.

Description

Merchandise display method, apparatus, equipment and medium based on image recognition
Technical field
The present invention relates to intelligent decision field more particularly to a kind of merchandise display method, apparatus based on image recognition, set Standby and medium.
Background technique
Currently, numerous businessmans can release new commodity during new product release or season as the time alternates, need The commodity of release are promoted, be bought with attracting clients.Usual businessman is that the commodity that will need to release pick out one A little commodity are shown, so that client's browsing is checked, but the hobby due to not knowing client, and chosen from commodity with personal eye It selects the commodity that the outstanding commodity of comparison are shown or intra-company personnel show needs to determine, to make The commodity that must be shown do not meet client's eye or demand, reduce the attraction for showing commodity, can not attract clients and buy into shop.
Summary of the invention
The embodiment of the present invention provides a kind of merchandise display method, apparatus, equipment and medium based on image recognition, to solve Show the inadequate problem of commodity attraction.
A kind of merchandise display method based on image recognition, comprising:
The target video data of commodity to be presented is obtained, the target video data includes at least two frame images to be recognized;
Recognition of face and cluster are carried out using Face datection model images to be recognized described at least two frames, obtain the mesh The corresponding customer quantity of video data and the corresponding image clustering set of each client are marked, each described image cluster set includes An at least frame images to be recognized;
If the customer quantity is greater than preset quantity, using preparatory trained micro- Expression Recognition model to each described Images to be recognized in image clustering set is identified, the single frames mood of each images to be recognized is obtained;
Based on the single frames mood of images to be recognized described in an at least frame, client corresponding with described image cluster set is obtained Target emotion;
The target emotion for gathering corresponding client according to the customer quantity and each described image cluster, obtains final feelings Thread;
According to the corresponding final mood of the commodity to be presented, intended display commodity are obtained.
A kind of commodity-exhibiting device based on image recognition, comprising:
Data acquisition module, for obtaining the target video data of commodity to be presented, the target video data includes extremely Few two frame images to be recognized;
Image clustering set obtains module, for being carried out using Face datection model images to be recognized described at least two frames Recognition of face and cluster obtain the corresponding customer quantity of the target video data and the corresponding image clustering collection of each client It closes, each described image cluster set includes an at least frame images to be recognized;
Single frames mood determining module, if being greater than preset quantity for the customer quantity, using trained micro- in advance Expression Recognition model identifies the images to be recognized in each described image cluster set, obtains each figure to be identified The single frames mood of picture;
Target emotion obtains module, for the single frames mood based on images to be recognized described in an at least frame, obtain with it is described The target emotion of the corresponding client of image clustering set;
Final mood obtains module, for gathering corresponding client according to the customer quantity and each described image cluster Target emotion, obtain final mood;
Intended display commodity obtain module, for obtaining target exhibition according to the corresponding final mood of the commodity to be presented Show commodity.
A kind of computer equipment, including memory, processor and storage are in the memory and can be in the processing The computer program run on device, the processor realize the above-mentioned commodity based on image recognition when executing the computer program The step of methods of exhibiting.
A kind of computer readable storage medium, the computer-readable recording medium storage have computer program, the meter Calculation machine program realizes the step of above-mentioned merchandise display method based on image recognition when being executed by processor.
Above-mentioned merchandise display method, apparatus, equipment and medium based on image recognition, by the mesh for obtaining commodity to be presented Video data is marked, recognition of face and cluster are carried out to images to be recognized in target video data using Face datection model, obtained The corresponding customer quantity of target video data and the corresponding image clustering set of each client, so as to subsequent according to a certain number of The target emotion of client determines intended display commodity, improves the accuracy rate of intended display commodity.If customer quantity is greater than present count Amount then uses preparatory trained micro- Expression Recognition model to identify the images to be recognized in each image clustering set, The single frames mood of each images to be recognized is obtained, to realize the mood of identification client.List based on an at least frame images to be recognized Frame mood obtains the target emotion of client corresponding with image clustering set, to determine whether the commodity to be presented are client institute Interested commodity.It is certain to get according to the target emotion of customer quantity and the corresponding client of each image clustering set The client of quantity obtains intended display according to the corresponding final mood of commodity to be presented to the final mood of the commodity to be presented Commodity improve the accuracy rate of intended display commodity, so that end article is the more interested commodity of most of clients.
Detailed description of the invention
In order to illustrate the technical solution of the embodiments of the present invention more clearly, below by institute in the description to the embodiment of the present invention Attached drawing to be used is needed to be briefly described, it should be apparent that, the accompanying drawings in the following description is only some implementations of the invention Example, for those of ordinary skill in the art, without any creative labor, can also be according to these attached drawings Obtain other attached drawings.
Fig. 1 is the application environment schematic diagram of the merchandise display method in one embodiment of the invention based on image recognition;
Fig. 2 is the flow chart of the merchandise display method in one embodiment of the invention based on image recognition;
Fig. 3 is the flow chart of the merchandise display method in one embodiment of the invention based on image recognition;
Fig. 4 is the flow chart of the merchandise display method in one embodiment of the invention based on image recognition;
Fig. 5 is the flow chart of the merchandise display method in one embodiment of the invention based on image recognition;
Fig. 6 is the flow chart of the merchandise display method in one embodiment of the invention based on image recognition;
Fig. 7 is the flow chart of the merchandise display method in one embodiment of the invention based on image recognition;
Fig. 8 is the functional block diagram of the commodity-exhibiting device in one embodiment of the invention based on image recognition;
Fig. 9 is a schematic diagram of computer equipment in one embodiment of the invention.
Specific embodiment
Below by the attached drawing in knot and the embodiment of the present invention, technical solution in the embodiment of the present invention carries out clear, complete Site preparation description, it is clear that described embodiments are some of the embodiments of the present invention, instead of all the embodiments.Based on this hair Embodiment in bright, every other implementation obtained by those of ordinary skill in the art without making creative efforts Example, shall fall within the protection scope of the present invention.
Merchandise display method provided in an embodiment of the present invention based on image recognition, can be applicable to the application environment such as Fig. 1 In, wherein client is communicated by network with server-side.The merchandise display method based on image recognition, which is applied, to be serviced On end, by carrying out analysis identification to the corresponding target video data of commodity to be presented, each visitor in target video data is obtained The corresponding mood in family, and client is determined to the final mood of the commodity to be presented, according to every according to the corresponding mood of each client The corresponding final mood of one commodity to be presented determines intended display commodity, so that intended display commodity are the quotient that client more pays close attention to Product improve the attraction of intended display commodity, are bought with attracting clients to displaying commodity.Wherein, client can with but not It is limited to various personal computers, laptop, smart phone, tablet computer and portable wearable device.Server-side can be with It is realized with the server-side cluster of the either multiple server-side compositions of independent server-side.
In one embodiment, it as shown in Fig. 2, providing a kind of merchandise display method based on image recognition, answers in this way It is illustrated, specifically comprises the following steps: for the server-side in Fig. 1
S10: obtaining the target video data of commodity to be presented, and target video data includes at least two frame images to be recognized.
Wherein, target video data, which refers to, screens to get initial video data corresponding with commodity to be presented Video data, specifically can be the video data for meeting some conditions.For example, meet the attribute information of commodity to be presented Video data.Attribute information may include Suitable Age and suitable gender.It is to be appreciated that passing through Suitable Age and suitable gender pair Collected initial video data corresponding with commodity to be presented is screened, the target video data got.Figure to be identified It seem the image referred to after being screened according to Suitable Age and suitable gender.Wherein, initial video data is video acquisition tool Collected video data corresponding with each commodity to be presented.
Specifically, a video acquisition tool, the video acquisition tool are configured for each commodity region to be presented in advance For acquiring image or video data, when video acquisition tool detect client appear in sampling instrument acquisition range it When interior, video acquisition tool automatic trigger, and image or video data acquiring are carried out to client.The video acquisition tool is specific For camera, the initial video data in the corresponding acquisition range of each commodity to be presented can be acquired by camera in real time.By Correspond to a commodity to be presented in each sampling instrument, each commodity region to be presented is acquired by each camera Initial video data, screened by the initial video data to commodity region to be presented, with get it is each with The corresponding target video data of commodity to be presented.Wherein, it is corresponding that commodity to be presented are carried in collected initial video data Commodity sign, to determine corresponding target video data subsequently through commodity sign.For example, video acquisition tool collects Initial video data include commodity sign A, then the initial video data be the corresponding initial video data of commodity sign A, Initial video data is screened, to get the corresponding target video data of commodity sign A.Wherein, commodity sign refers to For distinguishing the unique identification of different commodity to be presented.Optionally, commodity sign can be by number, letter, text or symbol Number at least one of composition.For example, commodity sign can be number either sequence number of commodity to be presented etc..
S20: recognition of face and cluster are carried out at least two frame images to be recognized using Face datection model, obtain target view For frequency according to corresponding customer quantity and the corresponding image clustering set of each client, each image clustering set includes an at least frame Images to be recognized.
Wherein, Face datection model refers to trained in advance for detecting whether each frame images to be recognized includes people's The model of facial area.Customer quantity refers in target video data, according to quantity determined by client's difference.Image clustering Set, which refers to, clusters the corresponding images to be recognized of same client, with the set of formation.
Specifically, server-side is connect with data bank network, and Face datection model, target video data are stored in database In include the corresponding images to be recognized of different clients, images to be recognized is identified using Face datection model, to get The facial image for including in images to be recognized, wherein facial image refers to the image of the human face region of client.Server-side will acquire To target video data in each images to be recognized be input to Face datection model, by Face datection model to target regard Frequency each images to be recognized in carries out recognition of face, determines whether each images to be recognized is facial image, if to be identified Image is facial image, then clusters identical facial image, i.e., to the corresponding images to be recognized of identical facial image It is clustered, to get the corresponding image clustering set of each client, according to image clustering set in target video data Quantity determines customer quantity in target video data.
More specifically, the face characteristic of the corresponding facial image of each images to be recognized is extracted using feature extraction algorithm, The corresponding face characteristic of each images to be recognized is subjected to characteristic similarity calculating, if characteristic similarity is greater than preset threshold, Illustrating that characteristic similarity is greater than the corresponding facial image of preset threshold is the corresponding facial image of same client, by same client couple The corresponding images to be recognized of the facial image answered is clustered, to obtain the corresponding image clustering set of each client, and according to The quantity of image clustering set determines the customer quantity in target video data.Wherein, preset when preset threshold to be used for Whether assessment similarity reaches the value for being determined as the face of same client.Feature extraction algorithm includes but is not limited to CNN (Convolutional Neural Network, convolutional Neural net) algorithm, for example, it is to be identified to extract people by CNN algorithm The face characteristic of the corresponding facial image of image.
S30: if customer quantity is greater than preset quantity, using preparatory trained micro- Expression Recognition model to each image Images to be recognized in cluster set is identified, the single frames mood of each images to be recognized is obtained.
Wherein, micro- Expression Recognition model refers to the local feature by capturing the client face in images to be recognized, and root According to each target face motor unit of face in local feature recognition images to be recognized, further according to the target face identified Motor unit determines the model of its mood.Single frames mood, which refers to, identifies images to be recognized by micro- Expression Recognition model, The mood determined according to the target face motor unit identified.
Wherein, micro- Expression Recognition model can be the neural network recognization model based on deep learning, can be and is based on dividing The local identification model of class can also be the local mood based on local binary patterns (Local Binary Pattern, LBP) Identification model.In the present embodiment, micro- Expression Recognition model is the local identification model based on classification, and micro- Expression Recognition model is preparatory It include each Facial action unit in training image data by collecting a large amount of training image data in advance when being trained Positive sample and Facial action unit negative sample, training image data are trained by sorting algorithm, obtain micro- expression Identification model.In the present embodiment, it can be and a large amount of training image data are trained by svm classifier algorithm, to obtain To SVM classifier corresponding with N number of Facial action unit.For example, it may be 39 Facial action units are 39 SVM points corresponding Class device is also possible to corresponding 54 SVM classifiers of 54 Facial action units, includes in the training image data being trained Different Facial action units positive sample and negative sample it is more, then the SVM classifier quantity got is more.It is understood that Ground forms micro- Expression Recognition model by N number of SVM classifier, and the SVM classifier got is more, then the micro- table formed The mood that feelings identification model is identified is more accurate.
Wherein, preset quantity is preset value, and the corresponding preset quantity of each commodity to be presented is identical, and server-side is sentenced Disconnected customer quantity out is greater than preset quantity, then using preparatory trained micro- Expression Recognition model in each image clustering set Images to be recognized identified, obtain the corresponding single frames mood of each images to be recognized.Using preparatory trained micro- expression Identification model carries out identification to the images to be recognized in each image clustering set and specifically comprises the following steps that server-side is first treated Identify that image carries out face critical point detection and feature extraction, to obtain corresponding local feature from images to be recognized, then will The corresponding local feature of images to be recognized is input in preparatory trained micro- Expression Recognition model.It is wrapped in micro- Expression Recognition model N number of SVM classifier is included, each SVM classifier identifies the corresponding local feature of an images to be recognized, passes through N number of SVM All local features of the input of classifier pair identify, obtain N number of SVM classifier output with the Facial action unit pair The Facial action unit is then determined as and the images to be recognized by the probability value answered when probability value is greater than predetermined probabilities threshold value Corresponding target face motor unit.Wherein, preset value when predetermined probabilities threshold value.Wherein, target face motor unit Refer to and images to be recognized is identified according to micro- Expression Recognition model, the Facial action unit that gets (Action Unit, AU)。
Include 54 SVM classifiers in micro- Expression Recognition model in the present embodiment, establishes Facial action unit number mapping Table, each Facial action unit are indicated with a prespecified number.For example, AU1 is that interior eyebrow raises up, AU2 is that outer eyebrow raises up, AU5 be upper eyelid raise up with AU26 be lower jaw open etc..Each Facial action unit trains corresponding SVM classifier.Example Such as, the probability value that interior eyebrow raises up is belonged to by the interior eyebrow local feature that corresponding SVM classifier may recognize that interior eyebrow raises up that raises up, Belong to the probability value etc. that outer eyebrow raises up by the outer eyebrow local feature that corresponding SVM classifier may recognize that outer eyebrow raises up that raises up. Probability value specifically can be the value between 0-1, if the probability value of output is 0.6, predetermined probabilities threshold value is 0.5, then probability value 0.6 is greater than predetermined probabilities threshold value 0.5, then by 0.6 corresponding Facial action unit, the target face as images to be recognized is acted Unit.
Specifically, server-side can determine the figure to be identified according to the corresponding target face motor unit of each images to be recognized The single frames mood of picture, i.e., according to the corresponding target face motor unit query assessment table of each images to be recognized, obtain with should be to Identify the corresponding single frames mood of image.In embodiment as above, local feature is identified by 54 SVM classifiers, really Fixed all target face motor units corresponding with images to be recognized, and according to all target face motor units, search assessment Table determines single frames mood corresponding with images to be recognized, to improve the accuracy for obtaining client's single frames mood.Wherein, table is assessed It is preconfigured table.One or more Facial action units form different moods, and it is corresponding to get each mood in advance The combination of each Facial action unit and corresponding mood are associated storage, to form assessment table by Facial action unit combination.
S40: the single frames mood based on an at least frame images to be recognized obtains the mesh of client corresponding with image clustering set Mark mood.
Wherein, target emotion refers to the feelings determined according to the corresponding single frames mood of each images to be recognized of image clustering set Thread.It is to be appreciated that being the corresponding images to be recognized of same client in image clustering set, pass through each to be identified of the client The corresponding single frames mood of image, determines the corresponding target emotion of the client.
Specifically, server-side obtains the corresponding single frames mood of each images to be recognized of image clustering set, to image clustering Gather the corresponding single frames mood of each images to be recognized to be analyzed, to obtain the target feelings of the corresponding client of image clustering set Thread.It is to be appreciated that judging whether the corresponding single frames mood of each images to be recognized of image clustering set is identical, if image clustering All single frames moods are identical in set, then using the single frames mood as target emotion.If at least two single frames moods are not identical, Then determine which kind of the corresponding quantity of single frames mood is most in image clustering set, using the at most corresponding single frames mood of quantity as mesh Mark mood.Specifically, the corresponding quantity of each single frames mood in statistical picture cluster set, by the corresponding single frames feelings of maximum quantity Target emotion of the thread as client corresponding with image clustering set.In the present embodiment, successively get and commodity pair to be presented The target emotion of the corresponding client of each image clustering set, that is, get each client and wait for this in the target video data answered Show the mood of commodity.For example, including 100 image clustering set in target video data corresponding with A commodity, obtain each The target emotion of the corresponding client of image clustering set gets 100 clients to the target emotion of A commodity.
S50: according to the target emotion of customer quantity and the corresponding client of each image clustering set, final mood is obtained.
Wherein, final mood, which refers to, carries out quantitative analysis by target emotion of each client to the commodity to be presented, obtains The mood corresponding with commodity to be presented got.
Specifically, server-side judges that all images are poly- according to the target emotion of the corresponding client of each image clustering set Whether the target emotion of the corresponding client of class set is identical, if the target emotion of the corresponding client of image clustering set is different, According to the target emotion of customer quantity and the corresponding client of each image clustering set, the corresponding number of each target emotion is counted Amount, using the maximum target emotion of quantity as final mood.For example, the corresponding customer quantity of the target video data of A commodity is 100, i.e. image clustering set is also 100, if the corresponding target emotion of 100 image clustering set, 50 are happiness, and 30 are Calmness, 20 are indifference, then the final mood by the maximum corresponding target emotion (i.e. happy) of quantity as A commodity.If number The maximum corresponding target emotion of amount is at least two, then determines final mood based on the quantity of the mood classification of target emotion.It is excellent Target emotion is the final mood of conduct of active mood by selection of land.For example, the corresponding client's number of the target video data of A commodity Amount be 100, i.e. image clustering set is also 100, if the corresponding target emotion of 100 image clustering set, 50 for happiness, 50 A is indifference, then is the happy final mood as the commodity to be presented by target emotion.If quantity maximum arranged side by side is corresponding Target emotion is at least two, and target emotion is negative emotions, since the intended display commodity finally shown are corresponding most Whole mood should be active mood, then any one target emotion may be selected as final mood.
S60: according to the corresponding final mood of commodity to be presented, intended display commodity are obtained.
Wherein, intended display commodity refer to according to the corresponding final mood of each commodity to be presented, from commodity to be presented Get the commodity that can be shown.
Specifically, server-side gets mesh from commodity to be presented according to the corresponding final mood of each commodity to be presented Mark shows that commodity specifically comprise the following steps:
(1) first determine whether the corresponding final mood of each commodity to be presented is the commodity that can be shown.Specifically, in advance The default mood that can be shown first is set, the corresponding final mood of each commodity to be presented is matched with default mood, It is that can be shown by the corresponding commodity to be presented of the final mood of successful match if final mood and default mood successful match Commodity.By the step, avoid target emotion and the unmatched commodity to be presented of default mood as intended display commodity.One As for, preset mood be active mood, then avoid using final mood be the corresponding commodity to be presented of negative emotions as mesh Mark shows commodity.For example, the corresponding final mood of A commodity to be presented is happiness, then most of client is to the commodity to be presented It is more interested, happiness is matched with default mood, if happy and default mood successful match, A commodity are determined as The commodity that can be shown.For another example the corresponding final mood of B commodity to be presented is to detest, is angry or disappointed, then it is most of Client does not like the commodity to be presented, and by the corresponding final mood of B commodity to be presented, it fails to match with default mood, then B not can be used as the commodity of displaying.
(2) quantity for the commodity that determination can be shown, if the quantity for the commodity that can be shown be greater than preset value, according to it is each can The corresponding final mood of the commodity of displaying inquires mood sequencing table, and the corresponding commodity to be presented of final mood are ranked up, are obtained Take the commodity to be presented of preset value as intended display commodity.Wherein, mood sequencing table is preset table, more positive feelings Before thread comes more, for example, being ranked up according to happy, happy, surprised, tranquil, detest, angry and disappointment etc..For example, can carry out The quantity of the commodity to be presented shown is 4, and preset value 3 then obtains the corresponding final feelings of commodity to be presented that can be shown Thread, the corresponding final mood of A commodity to be presented are happiness, and the corresponding final mood of B commodity to be presented is joy, C quotient to be presented The corresponding final mood of product be it is surprised, the corresponding final mood of D commodity to be presented be calmness, then being treated according to mood sequencing table Show that commodity are ranked up, as A, B, C and D, before the row of acquisition the commodity to be presented of preset value are arranged as intended display commodity The commodity to be presented of preceding 3 A, B and C are as intended display commodity.
Further, if according to mood sequencing table, occur simultaneously when the corresponding commodity to be presented of final mood are ranked up The case where column, first judges whether to get intended display commodity in preset value;If target exhibition cannot be got in preset value Show commodity, it is determined that quantity (the i.e. target emotion to match in the target video data of commodity to be presented with final mood arranged side by side Maximum quantity), the limited grade of final mood side by side is determined according to the quantity, obtains the intended display commodity of preset value.For example, The quantity for the commodity to be presented that can be shown is 4, preset value 3, then it is corresponding to obtain the commodity to be presented that can be shown Final mood, the corresponding final mood of A commodity to be presented are happiness, and the corresponding final mood of B commodity to be presented is joy, and C is waited for Show that the corresponding final mood of commodity is calmness, the corresponding final mood of D commodity to be presented is calmness, then sorting according to mood Table is ranked up commodity to be presented, as A, B, C and D, and C is arranged side by side with D, and intended display quotient cannot be got in preset value Product then obtain the quantity that final mood corresponding with C commodity to be presented matches in the target video data of C commodity to be presented, If the quantity is 50;In the target video data for determining D commodity to be presented again, final mood phase corresponding with D commodity to be presented Matched quantity, if the quantity is 60;It is then 60 corresponding D commodity to be presented by quantity prior to quantity is 50 corresponding C Commodity to be presented obtain the commodity to be presented of preset value as intended display commodity, i.e., the commodity to be presented of A, B and D are as mesh Mark shows commodity.By the step, final mood is avoided situation arranged side by side occur, to improve intended display commodity constant speed really Degree.
In step S10-S60, the target video data of commodity to be presented is obtained, using Face datection model to target video Images to be recognized carries out recognition of face and cluster in data, obtains the corresponding customer quantity of target video data and each client couple The image clustering set answered improves mesh so that the subsequent target emotion according to a certain number of clients determines intended display commodity Mark shows the accuracy rate of commodity.If customer quantity is greater than preset quantity, using preparatory trained micro- Expression Recognition model pair Images to be recognized in each image clustering set is identified, the single frames mood of each images to be recognized is obtained, and is known with realizing The mood of other client.Based on the single frames mood of an at least frame images to be recognized, obtain client's corresponding with image clustering set Target emotion, to determine whether the commodity to be presented are commodity interested to client.It is poly- according to customer quantity and each image The target emotion of the corresponding client of class set, to get a certain number of clients to the final mood of the commodity to be presented, according to According to the corresponding final mood of commodity to be presented, intended display commodity are obtained, the accuracy rate of intended display commodity are improved, so that target Commodity are the more interested commodity of most of clients.
In one embodiment, as shown in figure 3, in step S10, that is, the target video data of commodity to be presented, target are obtained Video data includes at least two frame images to be recognized, is specifically comprised the following steps:
S11: obtaining the initial video data of commodity to be presented, and initial video data includes at least two frame initial video figures Picture.
Wherein, initial video data refers to the video counts corresponding with each commodity to be presented of video acquisition tool acquisition According to.
Specifically, the corresponding initial video data of each commodity to be presented, initial video are acquired using video acquisition tool Include at least two frame initial video images in data, include the corresponding commodity sign of commodity to be presented in initial video data, It can determine initial video data corresponding with each commodity to be presented according to the commodity sign.By acquiring each commodity to be presented Corresponding initial video data analyzes initial video data so as to subsequent, and it is corresponding to get each commodity to be presented Final mood.
S12: obtaining the attribute information of commodity to be presented, and attribute information includes Suitable Age and suitable gender.
Wherein, Suitable Age referred to suitable for the commodity to be presented corresponding age.Suitable gender refers to that be suitable for should be to Show the corresponding gender of commodity.
Specifically, the corresponding attribute information of each commodity to be presented is stored in database.Server-side is according to quotient to be presented Product search database, obtain attribute information corresponding with each commodity to be presented.Wherein, in attribute information comprising Suitable Age and It is suitable for gender.For example, a certain commodity to be presented are dress ornament, Suitable Age is 20-24 years old in the corresponding attribute information of the dress ornament, is fitted The women that suitable gender is.For another example commodity to be presented are cosmetics, Suitable Age is in the corresponding attribute information of the cosmetics It is male etc. that 25-30 years old, which is suitable for gender,.Commodity to be presented are not done in the present embodiment specifically defined.
S13: screening initial video image using Suitable Age and suitable gender, obtains images to be recognized, is based on At least two frame images to be recognized form target video data.
Specifically, using preparatory trained classifier at least two frame initial video images in initial video data into Row identification obtains target age corresponding with each initial video image and target gender, and server-side is by target age and is suitable for Age is matched, and target gender is matched with suitable gender, successful match and and suitability will be carried out with Suitable Age Not carry out the initial video image of successful match be determined as images to be recognized, unsuccessful initial video image will be matched and deleted It removes, forms target video data based at least two frame images to be recognized.Wherein, target age refers to by trained point in advance Class device identifies initial video image, the accessed age.Target gender refers to through preparatory trained classifier pair Initial video image identified, accessed gender.
Step S11-S13 carries out initial video image according to the corresponding Suitable Age of commodity to be presented and suitable gender Screening, obtain images to be recognized, based at least two frame images to be recognized formed target video data so that target video data with Commodity to be presented more match, and improve the accuracy rate of intended display commodity.
In one embodiment, before step S3, i.e., using Suitable Age and suitable gender to initial video image into Before the step of row screening, the merchandise display method based on image recognition further include:
(1) at least two frame initial video images are handled using super-resolution technique, is obtained initial at least two frames The corresponding high-definition picture of video image, using high-definition picture as image to be determined.
Wherein, super-resolution technique (Super-Resolution, SR) refers to rebuilds from the low-resolution image got Corresponding high-definition picture out.In general, initial video image is low-resolution image in initial video data.Image to be determined Refer to and initial video image is converted into high-resolution image.
Specifically, server-side obtains initial video data, and initial video data includes at least two frame initial video images, just Beginning video image extracts the characteristic pattern of low-resolution spatial by ESPCN algorithm, by effective in the space low resolution (LR) Initial video image is amplified high-resolution from low resolution, final low resolution characteristic pattern is upgraded by sub-pix convolutional layer To high-resolution features figure, high-definition picture corresponding with each initial video image is obtained according to high-resolution features figure, Using high-definition picture as image to be determined.Wherein, the key concept of ESPCN algorithm is sub-pix convolutional layer (sub-pixel Convolutional layer), input is low-resolution image (i.e. initial video image), passes through two sub-pix convolutional layers After, obtained characteristic image size is as input picture, but feature channel is that (r is the target times magnification of image to r^2 Number).R^2 channel of each pixel is rearranged into the region of a r x r, corresponding to one in high-definition picture The sub-block of r x r size, so that size is that the characteristic image of r^2x H x W is rearranged into the height of 1x rH x rW size Resolution characteristics figure obtains high-definition picture corresponding with each frame initial video image according to high-resolution features figure, will High-definition picture is as image to be determined.
As shown in figure 4, being screened, being obtained to initial video image using Suitable Age and suitable gender in step S13 Images to be recognized is taken, is specifically comprised the following steps:
S131: identifying at least two frames image to be determined using preparatory trained classifier, obtain with it is each to Determine the corresponding target age of image and target gender.
Wherein, trained classifier includes gender sorter and character classification by age device in advance, passes through gender sorter and year Age classifier respectively identifies image to be determined, to get target age corresponding with image to be determined and Objective Not.Wherein, target gender, which refers to, identifies image to be determined by gender sorter, accessed gender.Target year Age, which refers to, identifies image to be determined by character classification by age device, the accessed age.
Wherein, when gender sorter and character classification by age device are trained, a large amount of training image data, training figure are first obtained Facial image as in data including all ages and classes and different sexes, carries out year for each facial image in training image data Training image data after mark are input to deep neural network by age and gender mark, and by deep neural network to mark Training image data after note are trained, since deep neural network includes at least two convolutional layers, by age predicted value and The age marked is compared, and with the weight and deviation of each layer in percentage regulation neural network, until model is restrained, obtains year Age classifier.Gender prediction's value and the gender marked are compared, with the weight of each layer in percentage regulation neural network and Deviation obtains gender sorter until model is restrained.
Specifically, image to be determined is identified using preparatory trained gender sorter, image to be determined is packet Image to be determined comprising client face is carried out face critical point detection and feature extraction, to obtain by the image of the face containing client Take face feature.Finally, the face feature of extraction is input in preparatory trained gender sorter, pass through gender sorter Face feature is identified, to obtain target gender corresponding with image to be determined.And the face feature that will be extracted is inputted Into preparatory trained character classification by age device, classified by character classification by age device to face feature, to obtain and figure to be determined As corresponding target age.Visitor is carried out to people's image to be determined by preparatory trained gender sorter and character classification by age device Family gender and age are estimated, and obtain accuracy to promote target gender and target age.
S132: target age is matched with Suitable Age, and target gender is matched with suitable gender.
Specifically, Suitable Age can be an age bracket, for example, 20-24 years old.Server-side is by target age and suitable year Whether age is matched, mainly judge the target age within the scope of Suitable Age.Be suitable for gender be women and two kinds of male, The target gender that will identify that is matched with suitable gender.
S133: if target age and Suitable Age successful match, and target gender and suitable gender successful match then will be with Target age and the corresponding image to be determined of target gender are as images to be recognized.
Specifically, server-side judge target age in Suitable Age segment limit, and target gender and suitable gender With success, then will image to be determined corresponding with target age and target gender as images to be recognized.
Step S131-S132 identifies at least two frames image to be determined using preparatory trained classifier, obtains Target age corresponding with each image to be determined and target gender, to realize through classifier to target age and target gender Determination, improve intended display commodity acquisition speed.If target age and Suitable Age successful match, and target gender and suitable Suitable gender successful match, then will image to be determined corresponding with target age and target gender as images to be recognized so that obtaining The images to be recognized got is corresponding with the attribute information of commodity to be presented, to improve the crowd's to match with product to be presented Attraction, so that the intended display commodity got are more accurate, improves intended display by analyzing images to be recognized The acquisition accuracy rate of commodity.
In one embodiment, as shown in figure 5, step S20, that is, use Face datection model at least two frame images to be recognized Recognition of face and cluster are carried out, the corresponding customer quantity of target video data and the corresponding image clustering collection of each client are obtained It closes, specifically comprises the following steps:
S21: recognition of face is carried out at least two frame images to be recognized using Face datection model, obtains each figure to be identified As corresponding facial image.
Specifically, server-side obtains target video data, using Face datection model to frame each in target video data Images to be recognized carries out recognition of face, gets the corresponding facial image of each images to be recognized in target video data.Wherein, Recognition of face refers to the image given for any one frame, use certain strategy to scan for it with determine in image whether Contain face.Further, Face datection model be in advance it is trained for detect each frame images to be recognized whether include The model of the facial area of people.Specifically, each frame images to be recognized is input in Face datection model by server-side, and detection is every Whether in one frame images to be recognized includes face, if in images to be recognized including face, is obtained every in target video data The corresponding facial image of one images to be recognized.
S22: clustering the corresponding facial image of images to be recognized, obtains at least two image clustering set, each Image clustering set includes an at least frame images to be recognized.
Specifically, the corresponding facial image of images to be recognized that server-side will acquire is clustered, and will include identical visitor The facial image at family is clustered, and at least two image clustering set are obtained, wherein includes at least in each image clustering set One frame images to be recognized.More specifically, extracting the people of the corresponding facial image of each images to be recognized using feature extraction algorithm The corresponding face characteristic of facial image is carried out characteristic similarity calculating by face feature, if characteristic similarity is greater than preset threshold, Illustrate the facial image for same client, and the corresponding images to be recognized of the facial image of same client is clustered, to obtain Take the corresponding image clustering set of each client.That is, the corresponding image clustering set of a client, each image clustering set In include an at least frame images to be recognized.
S23: according to the quantity of image clustering set, the corresponding customer quantity of target video data is obtained.
Specifically, the quantity for counting the corresponding image clustering set of each commodity to be presented, by the number of image clustering set Amount is as customer quantity corresponding with target video data.
Step S21-S23 carries out recognition of face at least two frame images to be recognized using Face datection model, obtains each The corresponding facial image of images to be recognized is avoided with realizing determine whether images to be recognized is facial image not comprising face Images to be recognized is clustered, and the acquisition speed of subsequent image cluster set is improved.Facial image corresponding to images to be recognized It is clustered, obtains at least two image clustering set, according to the quantity of image clustering set, it is corresponding to obtain target video data Customer quantity guarantee the accuracy that customer quantity obtains to realize the determination of customer quantity.
In one embodiment, the corresponding time identifier of each images to be recognized, time identifier refer to that collect this to be identified The image corresponding time.
As shown in fig. 6, clustering to the corresponding facial image of images to be recognized in step S22, at least two are obtained Image clustering set, specifically comprises the following steps:
S221: according to time identifier, using the facial image recognized for the first time at least two frame images to be recognized as benchmark Image.
Wherein, benchmark image refers to the facial image recognized for the first time from images to be recognized.
Specifically, server-side obtains the corresponding time identifier of at least two frame images to be recognized and is first determined according to time identifier The facial image recognized for the first time at least two frame images to be recognized, using the facial image as benchmark image.By determining base Quasi- image, to improve the acquisition speed of image clustering set.
S222: according to time identifier, successively using the spy of similarity algorithm calculating benchmark image and remaining images to be recognized Levy similarity.
Specifically, it according to time identifier, will be removed in benchmark image and at least two frame images to be recognized using similarity algorithm Remaining images to be recognized except benchmark image carries out characteristic similarity calculating, to obtain characteristic similarity.Wherein, similarity operator Method can be Euclidean distance algorithm, manhatton distance algorithm, Minkowski distance algorithm or cosine similarity algorithm Deng.It in the present embodiment, can using the characteristic similarity of cosine similarity algorithm calculating benchmark image and remaining images to be recognized Accelerate the acquisition speed of image clustering set, improves the acquisition efficiency of intended display commodity.
S223: if characteristic similarity be greater than preset threshold, by characteristic similarity be greater than preset threshold images to be recognized and Benchmark image belongs to same image clustering set.
Specifically, preset threshold is preset value, if server-side judges benchmark image and remaining images to be recognized Characteristic similarity be greater than preset threshold, then it is assumed that benchmark image and remaining images to be recognized successful match, benchmark image and surplus Remaining images to be recognized is the image of same client, then characteristic similarity is greater than to the images to be recognized and benchmark image of preset threshold Belong to same image clustering set.For example, the characteristic similarity corresponding with remaining images to be recognized 2 of benchmark image 1 is 80%, The characteristic similarity corresponding with remaining images to be recognized 3 of benchmark image 1 is 99%, preset threshold 90%, then benchmark image 1 Characteristic similarity corresponding with remaining images to be recognized 3 is greater than preset threshold, then by benchmark image 1 and remaining images to be recognized 3 Belong to same image clustering set.
S224: if characteristic similarity is not more than preset threshold, according to time identifier, characteristic similarity is not more than default threshold First in the remaining images to be recognized of value is updated to new benchmark image, repeats according to time identifier, successively uses The step of similarity algorithm calculating benchmark image and the characteristic similarity of remaining images to be recognized, until at least two frames figure to be identified As cluster set completion, at least two image clustering set are formed.
Specifically, server-side judges that the characteristic similarity of benchmark image and remaining images to be recognized is not more than default threshold Value, then it is assumed that it fails to match for benchmark image and remaining images to be recognized, the corresponding client of benchmark image and remaining images to be recognized Corresponding client is not same client, then according to time identifier, characteristic similarity is to be identified no more than the residue of preset threshold First in image is updated to new benchmark image.For example, the feature phase corresponding with remaining images to be recognized 2 of benchmark image 1 It is 80% like degree, preset threshold 90%, then the characteristic similarity corresponding with remaining images to be recognized 2 of benchmark image 1 is little In preset threshold, then according to time identifier, remaining images to be recognized 2 is updated to new benchmark image.It repeats according to the time The step of identifying, successively using similarity algorithm calculating benchmark image and the characteristic similarity of remaining images to be recognized, until extremely All images to be recognized cluster set are completed in few two frame images to be recognized, form at least two image clustering set.
Step S221-S224, according to time identifier, the facial image that will be recognized for the first time at least two frame images to be recognized As benchmark image, the characteristic similarity of similarity algorithm calculating benchmark image and remaining images to be recognized is successively used, so as to Determine whether benchmark image and remaining images to be recognized are same client.If characteristic similarity is greater than preset threshold, by feature phase It is greater than the images to be recognized of preset threshold like degree and benchmark image belongs to same image clustering set, realizes same client Images to be recognized clustered.If characteristic similarity is not more than preset threshold, according to time identifier, characteristic similarity is little First in the remaining images to be recognized of preset threshold is updated to new benchmark image, repeats according to time identifier, Successively using similarity algorithm calculating benchmark image and the characteristic similarity of remaining images to be recognized the step of, until at least two frames Images to be recognized cluster set is completed, and at least two image clustering set are formed, to realize the images to be recognized of same client It is clustered, so that each client of subsequent determination is to the target emotion of the commodity to be presented.
In implementing one, as shown in fig. 7, in step S30, i.e., using preparatory trained micro- Expression Recognition model to each Images to be recognized in image clustering set is identified, is obtained the single frames mood of each images to be recognized, is specifically included as follows Step:
S31: recognition of face is carried out to the images to be recognized in each image clustering set using face key point algorithm, is obtained Take face key point corresponding with each images to be recognized.
Wherein, face key point algorithm can be but not limited to Ensemble ofRegression Tress (abbreviation ERT) Algorithm, SIFT (scale-invariant feature transform) algorithm, SURF (Speeded Up Robust Features) algorithm, LBP (Local Binary Patterns) algorithm and HOG (Histogram ofOriented Gridients) algorithm.In the present embodiment, using ERT algorithm, people is carried out to the images to be recognized in each image clustering set Face identification, to get the corresponding face key point of each images to be recognized.Wherein, ERT algorithm is a kind of side based on recurrence Method, ERT algorithm are formulated as follows:Wherein,It is obtained for the t+1 times iteration The shape or coordinate of the characteristic point of images to be recognized, t indicate cascade serial number,For predict image characteristic point shape or Coordinate, I are the images to be recognized for returning device input, rtIndicate t grades of recurrence device, each recurrence device is by many regression trees (tree) it forms, by the available regression tree of training, it is crucial that the corresponding face of each images to be recognized is got by regression tree Point.
S32: feature extraction is carried out to the corresponding face key point of each images to be recognized using feature extraction algorithm, is obtained The corresponding local feature of face key point.
Wherein, feature extraction algorithm can be calculated with CNN (Convolutional Neural Network, convolutional Neural net) Method extracts the local feature of images to be recognized corresponding face key point position by CNN algorithm, specifically according to face action Local feature is extracted in the corresponding position of unit.Wherein, CNN algorithm is a kind of feedforward neural network, its artificial neuron can be with The surrounding cells in a part of coverage area are responded, image procossing can be rapidly and efficiently carried out.In the present embodiment, using preparatory Trained convolutional neural networks rapidly extract the corresponding local feature of face key point.
Specifically, by the corresponding face key point of each images to be recognized, convolution algorithm is carried out by several convolution kernels, Result after convolution is the corresponding local feature of face test point.Particular by formula (0≤m < M, 0≤n < N) carries out convolution algorithm, to get local feature.Wherein, y is the local feature of output, and x is one Size is the two dimension input variable of (M, N), i.e., is formed by the coordinate of L face key point, wijBe size be I*J convolution kernel, b For biasing, size M*N, activation primitive is indicated with f, and the face of input images to be recognized of each convolution kernel with upper one layer closes Key point carries out convolution operation, and each convolution kernel can have a corresponding local feature, and weight is shared in convolution kernel, number of parameters Reduce significantly, with greatly improving network training speed.
Further, face key point is input to after being identified in preset convolutional neural networks, available face The corresponding local feature of portion's motor unit.For example, AU1, AU2, AU5 and AU26, i.e., interior eyebrow raises up, outer eyebrow raises up, on upper eyelid It raises and opens corresponding local feature with lower jaw.It is crucial to face in images to be recognized using convolutional neural networks in the present embodiment The local feature of point extracts, determine target face motor unit according to local feature so as to subsequent, and according to recognizing Target face motor unit determines the mood of client.In this motion, relative to LBP-TOP operator, convolutional Neural net is used Faster, and accuracy of identification is higher for the arithmetic speed that network is identified.
S33: local feature is identified using preparatory trained classifier, obtains the corresponding mesh of each local feature Mark Facial action unit.
Specifically, local feature is known by each SVM classifier in preparatory trained micro- Expression Recognition model Not, wherein SVM classifier is identical as the quantity of identifiable Facial action unit, i.e., identifiable Facial action unit is 54 It is a, then trained SVM classifier is 54 in advance, by the way that local feature to be input in corresponding SVM classifier, obtain To probability value, probability value and the predetermined probabilities threshold value that will acquire are compared, and will be greater than the probability value pair of predetermined probabilities threshold value The Facial action unit answered is got corresponding with local feature as target face motor unit corresponding with local feature All target face motor units.
S34: being based on the corresponding target face motor unit of each local feature, searches assessment table, obtains each to be identified The single frames mood of image.
Wherein, assessment table is preconfigured table, and it is corresponding with mood that Facial action unit combination is stored in assessment table Relationship, for example, AU12, AU6 and AU7 are combined, corresponding mood is happy, AU9, AU10, AU17 and AU24, and corresponding mood is Detest.Server-side searches assessment table by the corresponding target face motor unit of each local feature, obtains dynamic with target face The combination to match as unit, using the corresponding mood of the combination as single frames mood corresponding with images to be recognized.
In step S31-S34, the images to be recognized in each image clustering set is carried out using face key point algorithm Recognition of face obtains face key point corresponding with each images to be recognized, provides technical assistance for subsequent extracted local feature, To improve the accuracy rate of local shape factor;Feature extraction is carried out to face key point using feature extraction algorithm, quickly to obtain The corresponding local feature of face key point is got, so that target face motor unit that subsequent extracted arrives is more accurate;Using pre- First trained classifier identifies local feature, single with the corresponding target face movement of each local feature of quick obtaining Member realizes the determination of target face motor unit.Based on the corresponding target face motor unit of each local feature, assessment is searched Table obtains the single frames mood of each images to be recognized, to determine the corresponding mood of each client according to single frames mood, according to every The corresponding mood of one client determines intended display commodity, to improve the accuracy rate of intended display commodity.
It should be understood that the size of the serial number of each step is not meant that the order of the execution order in above-described embodiment, each process Execution sequence should be determined by its function and internal logic, the implementation process without coping with the embodiment of the present invention constitutes any limit It is fixed.
In one embodiment, a kind of commodity-exhibiting device based on image recognition is provided, it should the commodity based on image recognition Show that the merchandise display method in device and above-described embodiment based on image recognition corresponds.As shown in figure 8, image should be based on The commodity-exhibiting device bag data of identification obtains module 10, image clustering set obtains module 20, single frames mood determining module 30, Target emotion obtains module 40, final mood obtains module 50 and intended display commodity obtain module 60.Each functional module is detailed It is described as follows:
Data acquisition module 10, for obtaining the target video data of commodity to be presented, target video data includes at least Two frame images to be recognized.
Image clustering set obtains module 20, for carrying out people at least two frame images to be recognized using Face datection model Face identification and cluster obtain the corresponding customer quantity of target video data and the corresponding image clustering set of each client, each Image clustering set includes an at least frame images to be recognized.
Single frames mood determining module 30, if being greater than preset quantity for customer quantity, using preparatory trained micro- table Feelings identification model identifies the images to be recognized in each image clustering set, obtains the single frames feelings of each images to be recognized Thread.
Target emotion obtains module 40, for the single frames mood based on an at least frame images to be recognized, obtains poly- with image The target emotion of the corresponding client of class set.
Final mood obtains module 50, for the target according to customer quantity and the corresponding client of each image clustering set Mood obtains final mood.
Intended display commodity obtain module 60, for obtaining intended display according to the corresponding final mood of commodity to be presented Commodity.
In one embodiment, data acquisition module 10, including initial video data acquiring unit 11, attribute information determine list Member 12 and target video data form unit 13.
Initial video data acquiring unit 11, for obtaining the initial video data of commodity to be presented, initial video data Including at least two frame initial video images.
Attribute information determination unit 12, for obtaining the attribute information of commodity to be presented, attribute information includes Suitable Age With suitable gender.
Target video data forms unit 13, for being sieved using Suitable Age and suitable gender to initial video image Choosing obtains images to be recognized, forms target video data based at least two frame images to be recognized.
In one embodiment, before target video data forms unit, the commodity-exhibiting device based on image recognition is also Including image resolution ratio converting unit.
Image resolution ratio converting unit, for using super-resolution technique at least two frame initial video images at Reason obtains high-definition picture corresponding at least two frame initial video images, using high-definition picture as image to be determined.
Target video data formed unit, including target age and target gender determine subelement, coupling subelement and to Identification image determines subelement.
Target age and target gender determine subelement, for being waited for really using preparatory trained classifier at least two frames Determine image to be identified, obtains target age corresponding with each image to be determined and target gender.
Coupling subelement, for target age to be matched with Suitable Age, and by target gender and suitable gender into Row matching.
Images to be recognized determines subelement, if for target age and Suitable Age successful match, and target gender and suitable Suitable gender successful match, then will image to be determined corresponding with target age and target gender as images to be recognized.
In one embodiment, image clustering set obtains module 20, including facial image acquiring unit, image clustering set Acquiring unit and customer quantity determination unit.
Facial image acquiring unit, for carrying out face knowledge at least two frame images to be recognized using Face datection model Not, the corresponding facial image of each images to be recognized is obtained.
Image clustering set acquiring unit obtains at least for clustering to the corresponding facial image of images to be recognized Two image clustering set, each image clustering set include an at least frame images to be recognized.
Customer quantity determination unit obtains the corresponding visitor of target video data for the quantity according to image clustering set Amount amount.
In one embodiment, the corresponding time identifier of each images to be recognized.
Image clustering set acquiring unit includes that benchmark image determines subelement, characteristic similarity computing unit, the first figure Gather as cluster and determines that subelement and the second image clustering set determine subelement.
Benchmark image determines subelement, for that will recognize for the first time at least two frame images to be recognized according to time identifier Facial image as benchmark image.
Characteristic similarity computing unit, for according to time identifier, successively using similarity algorithm calculating benchmark image with The characteristic similarity of remaining images to be recognized.
First image clustering set determines subelement, if being greater than preset threshold for characteristic similarity, by characteristic similarity Same image clustering set is belonged to greater than the images to be recognized of preset threshold and benchmark image.
Second image clustering set determines subelement, if being not more than preset threshold for characteristic similarity, is marked according to the time Know, characteristic similarity is updated to new benchmark image no more than first in the remaining images to be recognized of preset threshold, weight It is multiple to execute according to time identifier, successively using the characteristic similarity of similarity algorithm calculating benchmark image and remaining images to be recognized The step of, until at least two frame images to be recognized cluster set is completed, form at least two image clustering set.
In one embodiment, single frames mood determining module 30, including face key point acquiring unit, local shape factor list Member, target face motor unit acquiring unit, single frames mood acquiring unit.
Face key point acquiring unit, for using face key point algorithm to be identified in each image clustering set Image carries out recognition of face, obtains face key point corresponding with each images to be recognized.
Local shape factor unit, for using feature extraction algorithm to the corresponding face key point of each images to be recognized Feature extraction is carried out, the corresponding local feature of face key point is obtained.
Target face motor unit acquiring unit, for being known using preparatory trained classifier to local feature Not, the corresponding target face motor unit of each local feature is obtained.
Single frames mood acquiring unit searches assessment for being based on the corresponding target face motor unit of each local feature Table obtains the single frames mood of each images to be recognized.
Specific restriction about the commodity-exhibiting device based on image recognition may refer to know above for based on image The restriction of other merchandise display method, details are not described herein.Each mould in the above-mentioned commodity-exhibiting device based on image recognition Block can be fully or partially through software, hardware and its group and to realize.Above-mentioned each module can be embedded in the form of hardware or independence In processor in computer equipment, it can also be stored in a software form in the memory in computer equipment, in order to Processor, which calls, executes the corresponding operation of the above modules.
In one embodiment, a kind of computer equipment is provided, which can be server-side, internal junction Composition can be as shown in Figure 9.The computer equipment include by system bus connect processor, memory, network interface and Database.Wherein, the processor of the computer equipment is for providing calculating and control ability.The memory packet of the computer equipment Include non-volatile memory medium, built-in storage.The non-volatile memory medium is stored with operating system, computer program and data Library.The built-in storage provides environment for the operation of operating system and computer program in non-volatile memory medium.The calculating The database of machine equipment is for storing Face datection model and the attribute information of commodity to be presented etc..The network of the computer equipment Interface is used to communicate with external terminal by network connection.To realize that one kind is based on when the computer program is executed by processor The merchandise display method of image recognition.
In one embodiment, a kind of computer equipment is provided, including memory, processor and storage are on a memory simultaneously The computer program that can be run on a processor, processor is realized when executing computer program in above-described embodiment to be known based on image The step of other merchandise display method, for example, step S10 shown in Fig. 2 to step S60, alternatively, Fig. 3 is to step shown in Fig. 7 Suddenly.Processor realizes each mould in the commodity-exhibiting device in above-described embodiment based on image recognition when executing computer program Block/unit function, for example, function of the module 10 shown in Fig. 8 to module 50.To avoid repeating, details are not described herein again.
In one embodiment, a kind of computer readable storage medium is provided, computer program, computer are stored thereon with The merchandise display method based on image recognition in above method embodiment is realized when program is executed by processor, for example, shown in Fig. 2 Step S10 to step S60, alternatively, Fig. 3 is to step shown in Fig. 7.The computer program is realized above-mentioned when being executed by processor The function of each module/unit in commodity-exhibiting device in embodiment based on image recognition, for example, module 10 shown in Fig. 8 is to mould The function of block 60.To avoid repeating, details are not described herein again.
Those of ordinary skill in the art will appreciate that realizing all or part of the process in above-described embodiment method, being can be with Relevant hardware is instructed to complete by computer program, computer program can be stored in a non-volatile computer and can be read In storage medium, the computer program is when being executed, it may include such as the process of the embodiment of above-mentioned each method.Wherein, the application To any reference of memory, storage, database or other media used in provided each embodiment, may each comprise non- Volatibility and/or volatile memory.Nonvolatile memory may include read-only memory (ROM), programming ROM (PROM), Electrically programmable ROM (EPROM), electrically erasable ROM (EEPROM) or flash memory.Volatile memory may include arbitrary access Memory (RAM) or external cache.By way of illustration and not limitation, RAM is available in many forms, such as static RAM (SRAM), dynamic ram (DRAM), synchronous dram (SDRAM), double data rate sdram (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronization link (Synchlink) DRAM (SLDRAM), memory bus (RambuS) directly RAM (RDRAM), straight Connect memory bus dynamic ram (DRDRAM) and memory bus dynamic ram (RDRAM) etc..
It is apparent to those skilled in the art that for convenience of description and succinctly, only with above-mentioned each function Can unit, module division progress for example, in practical application, can according to need and by above-mentioned function distribution by different Functional unit, module are completed, i.e., the internal structure of device are divided into different functional unit or module, to complete above description All or part of function.
The above embodiments are merely illustrative of the technical solutions of the present invention, rather than its limitations;Although with reference to the foregoing embodiments Invention is explained in detail, those skilled in the art should understand that: it still can be to aforementioned each implementation Technical solution documented by example is modified or equivalent replacement of some of the technical features;And these modification or Replacement, the spirit and scope for technical solution of various embodiments of the present invention that it does not separate the essence of the corresponding technical solution should all include Within protection scope of the present invention.

Claims (10)

1. a kind of merchandise display method based on image recognition characterized by comprising
The target video data of commodity to be presented is obtained, the target video data includes at least two frame images to be recognized;
Recognition of face and cluster are carried out using Face datection model images to be recognized described at least two frames, obtain the target view For frequency according to corresponding customer quantity and the corresponding image clustering set of each client, each described image cluster set includes at least One frame images to be recognized;
If the customer quantity is greater than preset quantity, using preparatory trained micro- Expression Recognition model to each described image Images to be recognized in cluster set is identified, the single frames mood of each images to be recognized is obtained;
Based on the single frames mood of images to be recognized described in an at least frame, the mesh of client corresponding with described image cluster set is obtained Mark mood;
The target emotion for gathering corresponding client according to the customer quantity and each described image cluster, obtains final mood;
According to the corresponding final mood of the commodity to be presented, intended display commodity are obtained.
2. the merchandise display method based on image recognition as described in claim 1, which is characterized in that described to obtain quotient to be presented The target video data of product, the target video data include at least two frame images to be recognized, comprising:
The initial video data of commodity to be presented is obtained, the initial video data includes at least two frame initial video images;
The attribute information of the commodity to be presented is obtained, the attribute information includes Suitable Age and suitable gender;
The initial video image is screened using the Suitable Age and the suitable gender, obtains images to be recognized, Target video data is formed based on images to be recognized described at least two frames.
3. the merchandise display method based on image recognition as described in claim 1, which is characterized in that described using described suitable Before the step of suitable age and the suitable gender screen the initial video image, the quotient based on image recognition Product methods of exhibiting further include:
It is handled using super-resolution technique initial video image described at least two frames, it is initial described in acquisition and at least two frames The corresponding high-definition picture of video image, using the high-definition picture as image to be determined;
It is described that the initial video image is screened using the Suitable Age and the suitable gender, obtain figure to be identified Picture, comprising:
Identified using preparatory trained classifier image to be determined described at least two frames, obtain with it is each described to true Determine the corresponding target age of image and target gender;
The target age is matched with the Suitable Age, and the target gender and the suitable gender are carried out Match;
If the target age and the Suitable Age successful match, and the target gender is matched into the suitable gender Function, then will image to be determined corresponding with the target age and the target gender as images to be recognized.
4. the merchandise display method based on image recognition as described in claim 1, which is characterized in that described to use Face datection Model images to be recognized described at least two frames carries out recognition of face and cluster, obtains the corresponding client of the target video data Quantity and the corresponding image clustering set of each client, comprising:
Recognition of face is carried out using Face datection model images to be recognized described at least two frames, obtains each figure to be identified As corresponding facial image;
The corresponding facial image of the images to be recognized is clustered, at least two image clustering set, each image are obtained Cluster set includes an at least frame images to be recognized;
According to the quantity of described image cluster set, the corresponding customer quantity of the target video data is obtained.
5. the merchandise display method based on image recognition as claimed in claim 4, which is characterized in that each figure to be identified As a corresponding time identifier;
It is described that the corresponding facial image of the images to be recognized is clustered, obtain at least two image clustering set, comprising:
According to time identifier, using the facial image recognized for the first time in images to be recognized described at least two frames as benchmark image;
According to time identifier, it is similar to the remaining feature of images to be recognized that the benchmark image is successively calculated using similarity algorithm Degree;
If the characteristic similarity is greater than preset threshold, the characteristic similarity is greater than to images to be recognized and the institute of preset threshold It states benchmark image and belongs to same image clustering set;
If the characteristic similarity is not more than preset threshold, according to time identifier, the characteristic similarity is not more than default threshold First in the remaining images to be recognized of value is updated to new benchmark image, repeat it is described according to time identifier, successively The step of benchmark image and the characteristic similarity of remaining images to be recognized are calculated using similarity algorithm, until at least two frames The images to be recognized cluster set is completed, and at least two image clustering set are formed.
6. the merchandise display method based on image recognition as described in claim 1, which is characterized in that described using training in advance Good micro- Expression Recognition model identifies the images to be recognized in each described image cluster set, obtain it is each it is described to Identify the single frames mood of image, comprising:
Using face key point algorithm to each described image cluster set in images to be recognized carry out recognition of face, obtain with The corresponding face key point of each images to be recognized;
Feature extraction is carried out to the corresponding face key point of each images to be recognized using feature extraction algorithm, described in acquisition The corresponding local feature of face key point;
The local feature is identified using preparatory trained classifier, obtains the corresponding mesh of each local feature Mark Facial action unit;
Based on the corresponding target face motor unit of each local feature, assessment table is searched, is obtained each described to be identified The single frames mood of image.
7. a kind of commodity-exhibiting device based on image recognition characterized by comprising
Data acquisition module, for obtaining the target video data of commodity to be presented, the target video data includes at least two Frame images to be recognized;
Image clustering set obtains module, for carrying out face using Face datection model images to be recognized described at least two frames Identification and cluster, obtain the corresponding customer quantity of the target video data and the corresponding image clustering set of each client, often One described image cluster set includes an at least frame images to be recognized;
Single frames mood determining module, if being greater than preset quantity for the customer quantity, using preparatory trained micro- expression Identification model identifies the images to be recognized in each described image cluster set, obtains each images to be recognized Single frames mood;
Target emotion obtains module, for the single frames mood based on images to be recognized described in an at least frame, acquisition and described image Cluster gathers the target emotion of corresponding client;
Final mood obtains module, for gathering the mesh of corresponding client according to the customer quantity and each described image cluster Mood is marked, final mood is obtained;
Intended display commodity obtain module, for obtaining intended display quotient according to the corresponding final mood of the commodity to be presented Product.
8. as claimed in claim 7 based on the commodity-exhibiting device of image recognition, which is characterized in that the data acquisition mould Block, comprising:
Initial video data acquiring unit, for obtaining the initial video data of commodity to be presented, the initial video data packet Include at least two frame initial video images;
Attribute information determination unit, for obtaining the attribute information of the commodity to be presented, the attribute information includes suitable year Age and suitable gender;
Target video data forms unit, for using the Suitable Age and the suitable gender to the initial video image It is screened, obtains images to be recognized, target video data is formed based on images to be recognized described at least two frames.
9. a kind of computer equipment, including memory, processor and storage are in the memory and can be in the processor The computer program of upper operation, which is characterized in that the processor realized when executing the computer program as claim 1 to The step of merchandise display method described in 6 any one based on image recognition.
10. a kind of computer readable storage medium, the computer-readable recording medium storage has computer program, and feature exists In realization is as described in any one of claim 1 to 6 based on the commodity of image recognition when the computer program is executed by processor The step of methods of exhibiting.
CN201910042541.3A 2019-01-17 2019-01-17 Merchandise display method, apparatus, equipment and medium based on image recognition Pending CN109815873A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910042541.3A CN109815873A (en) 2019-01-17 2019-01-17 Merchandise display method, apparatus, equipment and medium based on image recognition
PCT/CN2019/120988 WO2020147430A1 (en) 2019-01-17 2019-11-26 Image identification-based product display method, device, apparatus, and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910042541.3A CN109815873A (en) 2019-01-17 2019-01-17 Merchandise display method, apparatus, equipment and medium based on image recognition

Publications (1)

Publication Number Publication Date
CN109815873A true CN109815873A (en) 2019-05-28

Family

ID=66604505

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910042541.3A Pending CN109815873A (en) 2019-01-17 2019-01-17 Merchandise display method, apparatus, equipment and medium based on image recognition

Country Status (2)

Country Link
CN (1) CN109815873A (en)
WO (1) WO2020147430A1 (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110598790A (en) * 2019-09-12 2019-12-20 北京达佳互联信息技术有限公司 Image identification method and device, electronic equipment and storage medium
CN110650306A (en) * 2019-09-03 2020-01-03 平安科技(深圳)有限公司 Method and device for adding expression in video chat, computer equipment and storage medium
CN111062786A (en) * 2019-12-25 2020-04-24 创新奇智(青岛)科技有限公司 Model updating method based on establishment of commodity appearance characteristic mapping table
CN111144241A (en) * 2019-12-13 2020-05-12 深圳奇迹智慧网络有限公司 Target identification method and device based on image verification and computer equipment
CN111291623A (en) * 2020-01-15 2020-06-16 浙江连信科技有限公司 Heart physiological characteristic prediction method and device based on face information
CN111310602A (en) * 2020-01-20 2020-06-19 北京正和恒基滨水生态环境治理股份有限公司 System and method for analyzing attention of exhibit based on emotion recognition
WO2020147430A1 (en) * 2019-01-17 2020-07-23 深圳壹账通智能科技有限公司 Image identification-based product display method, device, apparatus, and medium
CN111563503A (en) * 2020-05-09 2020-08-21 南宁市第三中学 Minority culture symbol identification method
CN112363624A (en) * 2020-11-16 2021-02-12 新之航传媒科技集团有限公司 Interactive exhibition hall system based on emotion analysis
CN113177603A (en) * 2021-05-12 2021-07-27 中移智行网络科技有限公司 Training method of classification model, video classification method and related equipment
CN113762156A (en) * 2021-09-08 2021-12-07 北京优酷科技有限公司 Viewing data processing method, device and storage medium
CN114153342A (en) * 2020-08-18 2022-03-08 深圳市万普拉斯科技有限公司 Visual information display method and device, computer equipment and storage medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112215653A (en) * 2020-10-12 2021-01-12 珠海格力电器股份有限公司 Commodity recommendation method, commodity recommendation device, server and storage medium
CN112434711B (en) * 2020-11-27 2023-10-13 杭州海康威视数字技术股份有限公司 Data management method and device and electronic equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106843028A (en) * 2016-12-04 2017-06-13 徐州乐泰机电科技有限公司 A kind of Intelligent clothes cabinet control method comprising advertisement and its intelligent screening
WO2018143630A1 (en) * 2017-02-01 2018-08-09 삼성전자 주식회사 Device and method for recommending product
CN108509941A (en) * 2018-04-20 2018-09-07 北京京东金融科技控股有限公司 Emotional information generation method and device
CN108858245A (en) * 2018-08-20 2018-11-23 深圳威琳懋生物科技有限公司 A kind of shopping guide robot
CN109048934A (en) * 2018-08-20 2018-12-21 深圳威琳懋生物科技有限公司 A kind of intelligent shopping guide robot system
CN109191190A (en) * 2018-08-20 2019-01-11 深圳威琳懋生物科技有限公司 A kind of control method and computer readable storage medium of shopping guide robot

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109815873A (en) * 2019-01-17 2019-05-28 深圳壹账通智能科技有限公司 Merchandise display method, apparatus, equipment and medium based on image recognition

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106843028A (en) * 2016-12-04 2017-06-13 徐州乐泰机电科技有限公司 A kind of Intelligent clothes cabinet control method comprising advertisement and its intelligent screening
WO2018143630A1 (en) * 2017-02-01 2018-08-09 삼성전자 주식회사 Device and method for recommending product
CN108509941A (en) * 2018-04-20 2018-09-07 北京京东金融科技控股有限公司 Emotional information generation method and device
CN108858245A (en) * 2018-08-20 2018-11-23 深圳威琳懋生物科技有限公司 A kind of shopping guide robot
CN109048934A (en) * 2018-08-20 2018-12-21 深圳威琳懋生物科技有限公司 A kind of intelligent shopping guide robot system
CN109191190A (en) * 2018-08-20 2019-01-11 深圳威琳懋生物科技有限公司 A kind of control method and computer readable storage medium of shopping guide robot

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020147430A1 (en) * 2019-01-17 2020-07-23 深圳壹账通智能科技有限公司 Image identification-based product display method, device, apparatus, and medium
CN110650306A (en) * 2019-09-03 2020-01-03 平安科技(深圳)有限公司 Method and device for adding expression in video chat, computer equipment and storage medium
CN110598790A (en) * 2019-09-12 2019-12-20 北京达佳互联信息技术有限公司 Image identification method and device, electronic equipment and storage medium
CN111144241A (en) * 2019-12-13 2020-05-12 深圳奇迹智慧网络有限公司 Target identification method and device based on image verification and computer equipment
CN111062786A (en) * 2019-12-25 2020-04-24 创新奇智(青岛)科技有限公司 Model updating method based on establishment of commodity appearance characteristic mapping table
CN111291623A (en) * 2020-01-15 2020-06-16 浙江连信科技有限公司 Heart physiological characteristic prediction method and device based on face information
CN111310602A (en) * 2020-01-20 2020-06-19 北京正和恒基滨水生态环境治理股份有限公司 System and method for analyzing attention of exhibit based on emotion recognition
CN111563503A (en) * 2020-05-09 2020-08-21 南宁市第三中学 Minority culture symbol identification method
CN114153342A (en) * 2020-08-18 2022-03-08 深圳市万普拉斯科技有限公司 Visual information display method and device, computer equipment and storage medium
CN112363624A (en) * 2020-11-16 2021-02-12 新之航传媒科技集团有限公司 Interactive exhibition hall system based on emotion analysis
CN112363624B (en) * 2020-11-16 2022-09-09 新之航传媒科技集团有限公司 Interactive exhibition hall system based on emotion analysis
CN113177603A (en) * 2021-05-12 2021-07-27 中移智行网络科技有限公司 Training method of classification model, video classification method and related equipment
CN113177603B (en) * 2021-05-12 2022-05-06 中移智行网络科技有限公司 Training method of classification model, video classification method and related equipment
CN113762156A (en) * 2021-09-08 2021-12-07 北京优酷科技有限公司 Viewing data processing method, device and storage medium
CN113762156B (en) * 2021-09-08 2023-10-24 北京优酷科技有限公司 Video data processing method, device and storage medium

Also Published As

Publication number Publication date
WO2020147430A1 (en) 2020-07-23

Similar Documents

Publication Publication Date Title
CN109815873A (en) Merchandise display method, apparatus, equipment and medium based on image recognition
Pérez-Borrero et al. A fast and accurate deep learning method for strawberry instance segmentation
US20160350336A1 (en) Automated image searching, exploration and discovery
Othmani et al. Age estimation from faces using deep learning: A comparative analysis
WO2016190814A1 (en) Method and system for facial recognition
Anandhakrishnan et al. Deep Convolutional Neural Networks for image based tomato leaf disease detection
CN108960167A (en) Hair style recognition methods, device, computer readable storage medium and computer equipment
CN110413825B (en) Street-clapping recommendation system oriented to fashion electronic commerce
Karanam et al. Study on image processing using deep learning techniques
CN110580446A (en) Behavior semantic subdivision understanding method, system, computer device and medium
Biswas et al. Recognition of local birds using different cnn architectures with transfer learning
Benhabiles et al. Deep learning based detection of hair loss levels from facial images
Bondre et al. Review on leaf diseases detection using deep learning
Guehairia et al. Deep random forest for facial age estimation based on face images
CN110956157A (en) Deep learning remote sensing image target detection method and device based on candidate frame selection
Kukreja et al. Early Detection and Prevention of Mustard Downy Mildew Disease using a Hybrid CNN-LSTM Model
Rahaman et al. A deep learning based smartphone application for detecting mango diseases and pesticide suggestions
CN114239754A (en) Pedestrian attribute identification method and system based on attribute feature learning decoupling
Raja et al. A novel fuzzy-based modified GAN and faster RCNN for classification of banana leaf disease
Bashir et al. A comprehensive review on apple leaf disease detection
Parameshachari et al. Plant Disease Detection and Classification Using Transfer Learning Inception Technique
Milke et al. Development of a coffee wilt disease identification model using deep learning
Hall An Automated Approach to Bee Identi cation from Wing Venation
Park et al. Clothing classification using CNN and shopping mall search system
Chang et al. Research on classification and recognition method of plant leaves based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20190528