CN114547357B - Furniture model retrieval method based on voice and sketch - Google Patents

Furniture model retrieval method based on voice and sketch Download PDF

Info

Publication number
CN114547357B
CN114547357B CN202210043225.XA CN202210043225A CN114547357B CN 114547357 B CN114547357 B CN 114547357B CN 202210043225 A CN202210043225 A CN 202210043225A CN 114547357 B CN114547357 B CN 114547357B
Authority
CN
China
Prior art keywords
sketch
model
dimensional
voice
features
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210043225.XA
Other languages
Chinese (zh)
Other versions
CN114547357A (en
Inventor
武仲科
王醒策
赵禹铭
徐墅
王虎镇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Normal University
Original Assignee
Beijing Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Normal University filed Critical Beijing Normal University
Priority to CN202210043225.XA priority Critical patent/CN114547357B/en
Publication of CN114547357A publication Critical patent/CN114547357A/en
Application granted granted Critical
Publication of CN114547357B publication Critical patent/CN114547357B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • G06F16/532Query formulation, e.g. graphical querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/5866Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, manually generated location and time information
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Databases & Information Systems (AREA)
  • Library & Information Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Probability & Statistics with Applications (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention belongs to the technical field of furniture model retrieval in computer aided design, and particularly relates to a furniture model retrieval method based on voice and sketch, which comprises the following steps: s1, constructing a database based on a neural network, standard keyword description and model description semantic features; s2, inputting a two-dimensional sketch into a database to obtain a sketch retrieval result; s3, inputting voice into a database to obtain a voice retrieval result; s4, obtaining a final search result through a voting mode based on the sketch search result and the voice search result. The quick and accurate furniture model retrieval method disclosed by the invention saves a great deal of design cost and time expenditure, so that the utilization rate and the use value of the furniture model are improved.

Description

Furniture model retrieval method based on voice and sketch
Technical Field
The invention belongs to the technical field of furniture model retrieval in computer aided design, and particularly relates to a furniture model retrieval method based on voice and sketch.
Background
With the development of computer-aided designs, furniture designs increasingly rely on computer-aided designs. In the design of home scenes, the retrieval of furniture models is a critical issue. Due to the wide variety and number of furniture models, home designers need quick and accurate methods to retrieve furniture models. The furniture retrieval task mainly comprises two aspects of extracting model features and establishing a relationship between input and the model features.
The following search methods are generally used: text-based retrieval and sketch-based retrieval. The method based on text retrieval needs to carry out a large number of manual labels on the model, stores the model characteristics in a manual label mode, and compares the model characteristics with input text to obtain a retrieval result. And the sketch-based retrieval is performed by drawing a sketch of a model to be retrieved by a user, extracting the sketch features by the system, and calculating the similarity between the sketch features and the model features to obtain a retrieval result. The method has the advantages of no need of a large amount of manual labeling, and lower accuracy compared with the traditional text retrieval.
In text-based retrieval, due to the ambiguity of the input, a complete match with the model tag key is typically not possible, requiring text features to be expressed in a semantic-based manner. Word vector embedding is a semantic expression technique, and there are two types of word vector embedding methods: fixed representation and dynamic representation. The fixed representation method comprises the following steps: word2Vec, gloVe, fastText, etc., the dynamic representation method is: ELMo, bRT, etc. In sketch-based retrieval, two general categories can be distinguished: and extracting three-dimensional model features based on projection, and directly extracting three-dimensional model global features.
For example, the Chinese patent application number is: CN201810597066.1 discloses a three-dimensional model retrieval method based on sketch, which processes the three-dimensional model into a form of a plurality of view angle screenshots, then uses different contour extraction operators to process the three-dimensional model into sketches with different styles, and obtains a sketch dataset, and the dataset is used in training and testing processes after being provided with labels; the invention relates to a method for searching a furniture model based on voice and sketch, which is characterized in that a hierarchical network method is added on the basis of a convolutional neural network, 1 large classification network is used for training large classifications, the input sketch is classified into 40 large classes, 40 small classification networks respectively train specific differences in each class of models, the input sketch is classified into one specific three-dimensional model in a certain large class, the three-dimensional model can be searched with high accuracy, the information redundancy is small, an angle matrix calculated by using sampling contour points is used as the input of the convolutional neural network, the characteristics of the sketch are more consistent, and the searching accuracy is high, but a furniture model searching method based on voice and sketch is not proposed.
Disclosure of Invention
In order to solve the problems, the invention provides a furniture model retrieval method based on voice and sketch;
S1, constructing a database based on a neural network, standard keyword description and model description semantic features;
S2, inputting a two-dimensional sketch into a database to obtain a sketch retrieval result;
s3, inputting voice into a database to obtain a voice retrieval result;
s4, obtaining a final search result through a voting mode based on the sketch search result and the voice search result.
Further, the constructing the database based on the neural network, the standard keyword description and the model description semantic features in S1 includes:
s101, constructing a pseudo-twin neural network for calculating similarity calculation of two-dimensional sketch features and three-dimensional model features, taking the three-dimensional model features and the two-dimensional sketch features as input values of the pseudo-twin neural network, and extracting features of three-dimensional models in a database by adopting an LD-SIFT descriptor mode to obtain three-dimensional model features as input values of the pseudo-twin neural network;
S102, constructing standard keyword descriptions for standardizing keyword descriptions of input search texts, acquiring texts corresponding to all three-dimensional models in advance, inputting the texts into a word2Vec pre-training model for word vector embedding, and obtaining vector representations of all model description keywords as all standard keyword descriptions;
S103, constructing all model description semantic features corresponding to all standard keyword descriptions, acquiring standard keyword descriptions corresponding to each three-dimensional model, and calculating IDF weights corresponding to each three-dimensional model by adopting IDF through the standard keyword descriptions to serve as all model description semantic features;
S104, constructing a database by constructing a pseudo twin neural network, standard keyword description and model description semantic features.
Further, the step S2 of inputting the two-dimensional sketch into the database to obtain a sketch retrieval result includes:
s201, drawing a two-dimensional sketch of a model to be searched by a user, inputting the two-dimensional sketch into a trained VGG-19 model, and acquiring a 4096-dimensional vector as an input sketch characteristic through an fc-7 layer of the trained VGG-19 model;
s202, inputting the input sketch features and the three-dimensional model features into a pseudo-twin neural network, and obtaining the similarity of the input sketch features and the three-dimensional model features as sketch retrieval results.
Further, in S202, the similarity between the input sketch feature and the three-dimensional model feature is calculated as the following formula (1):
In the method, in the process of the invention, Representing input sketch features and three-dimensional model feature similarities,A contrast loss function is represented and is used,Represented as a sample feature 1,The characteristic is expressed as a sample characteristic 2, W is a neural network parameter, Y is expressed as whether samples are matched, and when Y is 1, the samples are matched; at 0, the samples do not match.
Further, in S202, the contrast loss function is calculated as follows (2):
Where N is the number of samples, D W is the euclidean distance between sample X 1 and sample X 2, m is the set upper distance threshold, and max (m-D w, 0) is the maximum value of m-D w, 0.
Further, in S202, the euclidean distance is calculated as follows (3):
In the method, in the process of the invention, the term "is expressed as the modulo length of the vector, X 1 represents sample 1 and X 2 represents sample 2.
Further, in S3, the step of inputting the voice into a pre-constructed database to obtain a voice search result includes:
s301, converting input voice into text data by adopting a voice recognition algorithm, performing word segmentation and word removal operation on the text data to obtain a keyword description sequence, and inputting the keyword description sequence into a word2Vec pre-training model to perform word vector embedding to obtain a keyword vector;
S302, carrying out similarity calculation on the keyword vector and standard keyword description corresponding to the database, converting the keyword description sequence into a standard keyword sequence through an optimal similarity principle, and carrying out IDF feature extraction on the standard keyword sequence to obtain IDF weight corresponding to the standard keyword sequence;
s303, determining a three-dimensional model based on the corresponding standard keyword description, and selecting corresponding model description semantic features through the three-dimensional model to serve as IDF weights corresponding to the three-dimensional model in a database;
S304, obtaining an inner product corresponding to the three-dimensional model through the IDF weight corresponding to the standard keyword sequence and the IDF weight corresponding to the three-dimensional model in the database, and taking the inner product as the similarity between the output voice and the three-dimensional model in the pre-built database.
Further, in S303, the IDF weight is calculated as follows (4):
Where W tag represents the weight, N model represents the number of three-dimensional models, and N tag represents the number of labels corresponding to the standard keyword sequence.
Further, in S4, the obtaining the final search result based on the sketch search result and the voice search result through the voting method includes:
S401, respectively acquiring the similarity of all three-dimensional models according to the input two-dimensional sketch and the input voice, respectively sequencing all the three-dimensional models from large to small according to the similarity, and distributing weights;
s402, voting the two search results according to the weights to obtain total weights, and sorting the two search results according to the total weights from high to low to obtain a final search result.
Compared with the prior art, the invention has the following beneficial effects:
1. The furniture model retrieval method based on the voice and the sketch overcomes the defects brought by a single retrieval method, combines the advantages of the sketch retrieval method and the voice retrieval method, achieves higher retrieval accuracy under the condition of insufficient labeling of text keywords, and meets higher retrieval quality requirements.
2. The furniture model retrieval method based on the voice and the sketch designs a quick and accurate furniture model retrieval method, and can save a great deal of design cost and time expenditure, thereby improving the utilization rate and the use value of the furniture model.
3. According to the furniture model retrieval method based on the voice and the sketch, the voice information and the sketch information are effectively extracted, the similarity comparison of heterogeneous data is realized through fusion, in the sketch retrieval, the sketch features and the model features are extracted by using VGG19 and LD-SIFT respectively, the registration of the features is carried out in a pseudo twin neural network, the two features are mapped into the same vector space, and the retrieval is carried out through calculating the similarity; in the retrieval based on voice, a Word2vec model is trained by using furniture related corpus, a semantic relation library of furniture keywords is established, the research is carried out by converting voice and text, extracting the furniture description keywords in the text, constructing keyword semantic relation, carrying out standard keyword conversion, giving weight to each keyword by using inverse document frequency, converting each furniture model description and query into vectors, realizing retrieval by similarity calculation of the vectors, and finally, merging two independent retrieval results by using a voting method, thereby realizing a multi-mode combined retrieval mode.
Drawings
FIG. 1 is a block flow diagram of a furniture model retrieval method based on voice and sketch according to the invention;
fig. 2 is a schematic flow chart of the furniture model searching method based on voice and sketch according to the invention, wherein the final searching result is obtained by voting based on the sketch searching result and the voice searching result.
Detailed Description
The invention is described in further detail below with reference to the drawings and the detailed description.
The invention provides a joint search method based on the existing model search method based on voice and the model search method based on sketch, and simultaneously refers to the results of the two search modes of the sketch search method and the voice search method, so that the search result is more accurate and the search mode is more efficient. As shown in fig. 1 and 2, the furniture model retrieval method based on voice and sketch includes:
s1, constructing a database based on a neural network, standard keyword description and model description semantic features:
s101, constructing a pseudo-twin neural network for calculating similarity calculation of two-dimensional sketch features and three-dimensional model features, taking the three-dimensional model features and the two-dimensional sketch features as input values of the pseudo-twin neural network, and extracting features of three-dimensional models in a database by adopting an LD-SIFT descriptor mode to obtain three-dimensional model features as input values of the pseudo-twin neural network;
S102, constructing standard keyword descriptions for standardizing keyword descriptions of input search texts, acquiring texts corresponding to all three-dimensional models in advance, inputting the texts into a word2Vec pre-training model for word vector embedding, and obtaining vector representations of all model description keywords as all standard keyword descriptions;
S103, constructing all model description semantic features corresponding to all standard keyword descriptions, acquiring standard keyword descriptions corresponding to each three-dimensional model, and calculating IDF weights corresponding to each three-dimensional model by adopting IDF through the standard keyword descriptions to serve as all model description semantic features;
S104, constructing a database by constructing a pseudo twin neural network, standard keyword description and model description semantic features.
The construction of the database mainly comprises 3 parts, namely, the construction of a neural network for calculating the similarity between a two-dimensional sketch and a three-dimensional model, the construction of standard keyword description and the construction of model description semantic features:
1) And constructing a neural network for calculating the similarity between the two-dimensional sketch and the three-dimensional model, and constructing the neural network for calculating the similarity between the two, wherein the characteristics of the two are extracted firstly. Extracting the features of the two-dimensional sketch by adopting VGG-19, extracting the fc-7 layer of a pre-trained VGG-19 model as the features of the input two-dimensional sketch, extracting the features of the three-dimensional model by adopting LD-SIFT descriptors, and constructing a pseudo-twin neural network on feature similarity calculation to train the features of the three-dimensional model and the features of the two-dimensional sketch so as to obtain a pseudo-twin neural network capable of calculating the similarity of the sketch and the model, wherein a loss function adopts a contrast loss function as an index of similarity measurement;
2) Constructing standard keyword descriptions, namely embedding word vectors into original model feature descriptions in a database through a word2Vec pre-training model to obtain vector representations of all model description keywords, wherein the vector representations are used for normalizing the keyword descriptions of the input retrieval text items;
3) Constructing semantic features of model description, constructing semantic feature description of each model by adopting IDF (inverse document frequency) for keyword description of all models in a database, calculating inverse document frequency of each keyword, and constructing the document frequency of the keyword which corresponds to each model as a vector which is the semantic feature of the model;
s2, inputting a two-dimensional sketch into a database to obtain a sketch retrieval result:
s201, drawing a two-dimensional sketch of a model to be searched by a user, inputting the two-dimensional sketch into a trained VGG-19 model, and acquiring a 4096-dimensional vector as an input sketch characteristic through an fc-7 layer of the trained VGG-19 model;
S202, inputting the input sketch features and the three-dimensional model features into a pseudo-twin neural network, and obtaining the similarity of the input sketch features and the three-dimensional model features as sketch retrieval results:
The similarity between the input sketch features and the three-dimensional model features is calculated as follows (1):
In the method, in the process of the invention, Representing input sketch features and three-dimensional model feature similarities,A contrast loss function is represented and is used,Represented as a sample feature 1,The characteristic is expressed as a sample characteristic 2, W is a neural network parameter, Y is expressed as whether samples are matched, and when Y is 1, the samples are matched; at 0, the samples do not match;
the contrast loss function is calculated according to the following formula (2):
Wherein N is the number of samples, D W is the Euclidean distance between the sample X 1 and the sample X 2, m is the set distance upper threshold, and max (m-D w, 0) is the maximum value of m-D w, 0;
the euclidean distance is calculated according to the following formula (3):
In the method, in the process of the invention, the term "is expressed as the modulo length of the vector, X 1 represents sample 1, and X 2 represents sample 2;
s3, inputting voice into a database to obtain a voice retrieval result:
s301, converting input voice into text data by adopting a voice recognition algorithm, performing word segmentation and word removal operation on the text data to obtain a keyword description sequence, and inputting the keyword description sequence into a word2Vec pre-training model to perform word vector embedding to obtain a keyword vector;
S302, carrying out similarity calculation on the keyword vector and standard keyword description corresponding to the database, converting the keyword description sequence into a standard keyword sequence through an optimal similarity principle, and carrying out IDF feature extraction on the standard keyword sequence to obtain IDF weight corresponding to the standard keyword sequence;
s303, determining a three-dimensional model based on the corresponding standard keyword description, and selecting corresponding model description semantic features through the three-dimensional model to serve as IDF weights corresponding to the three-dimensional model in a database;
S304, obtaining an inner product corresponding to the three-dimensional model through the IDF weight corresponding to the standard keyword sequence and the IDF weight corresponding to the three-dimensional model in the database, and taking the inner product as the similarity between the output voice and the three-dimensional model in the pre-built database:
The IDF weight is calculated as follows (4):
Wherein W tag represents weight, N model represents three-dimensional model number, and N tag represents label number corresponding to standard keyword sequence;
s4, obtaining a final search result through a voting mode based on the sketch search result and the voice search result:
S401, respectively acquiring the similarity of all three-dimensional models according to the input two-dimensional sketch and the input voice, respectively sequencing all the three-dimensional models from large to small according to the similarity, and distributing weights;
s402, voting the two search results according to the weights to obtain total weights, and sorting the two search results according to the total weights from high to low to obtain a final search result.
In step 4, the two search modes are combined by voting, different voting weights are allocated to different models according to the ranking of the search results, weights are respectively allocated from the first to the last of the search results, the two search results are voted according to the weights to obtain total weights, and finally the final search results are ranked according to the total weights from high to low.
The above description is only specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions easily come within the scope of the present invention as those skilled in the art can easily come within the scope of the present invention defined by the appended claims.

Claims (8)

1. A furniture model retrieval method based on voice and sketch is characterized by comprising the following steps:
S1, constructing a database based on a neural network, standard keyword description and model description semantic features;
s1, constructing a database based on a neural network, standard keyword description and model description semantic features, wherein the database comprises the following steps:
s101, constructing a pseudo-twin neural network for calculating similarity calculation of two-dimensional sketch features and three-dimensional model features, taking the three-dimensional model features and the two-dimensional sketch features as input values of the pseudo-twin neural network, and extracting features of three-dimensional models in a database by adopting an LD-SIFT descriptor mode to obtain three-dimensional model features as input values of the pseudo-twin neural network;
S102, constructing standard keyword descriptions for standardizing keyword descriptions of input search texts, acquiring texts corresponding to all three-dimensional models in advance, inputting the texts into a word2Vec pre-training model for word vector embedding, and obtaining vector representations of all model description keywords as all standard keyword descriptions;
S103, constructing all model description semantic features corresponding to all standard keyword descriptions, acquiring standard keyword descriptions corresponding to each three-dimensional model, and calculating IDF weights corresponding to each three-dimensional model by adopting IDF through the standard keyword descriptions to serve as all model description semantic features;
S104, constructing a database by constructing a pseudo twin neural network, standard keyword description and model description semantic features;
S2, inputting a two-dimensional sketch into a database to obtain a sketch retrieval result;
s3, inputting voice into a database to obtain a voice retrieval result;
s4, obtaining a final search result through a voting mode based on the sketch search result and the voice search result.
2. The method for retrieving a furniture model based on voice and sketch according to claim 1, wherein S2 of inputting the two-dimensional sketch into a database to obtain a sketch retrieval result comprises:
s201, drawing a two-dimensional sketch of a model to be searched by a user, inputting the two-dimensional sketch into a trained VGG-19 model, and acquiring a 4096-dimensional vector as an input sketch characteristic through an fc-7 layer of the trained VGG-19 model;
s202, inputting the input sketch features and the three-dimensional model features into a pseudo-twin neural network, and obtaining the similarity of the input sketch features and the three-dimensional model features as sketch retrieval results.
3. The furniture model retrieval method based on voice and sketch according to claim 2, wherein the calculation of the similarity between the input sketch features and the three-dimensional model features in S202 is as follows formula (1):
In the method, in the process of the invention, Representing input sketch features and three-dimensional model feature similarities,A contrast loss function is represented and is used,Represented as a sample feature 1,The characteristic is expressed as a sample characteristic 2, W is a neural network parameter, Y is expressed as whether samples are matched, and when Y is 1, the samples are matched; at 0, the samples do not match.
4. A furniture model retrieval method based on speech and sketch according to claim 3, characterized in that the calculation of the contrast loss function is as follows:
Where N is the number of samples, D W is the euclidean distance between sample X 1 and sample X 2, m is the set upper distance threshold, and max (m-D w, 0) is the maximum value of m-D w, 0.
5. The furniture model retrieval method based on voice and sketch according to claim 4, wherein the calculation of the euclidean distance is as follows:
In the formula, X 1 represents sample 1, and X 2 represents sample 2, expressed as a modulus of vector.
6. The furniture model retrieval method based on voice and sketch according to claim 1, wherein S3, obtaining a voice retrieval result by inputting voice into a pre-built database includes:
s301, converting input voice into text data by adopting a voice recognition algorithm, performing word segmentation and word removal operation on the text data to obtain a keyword description sequence, and inputting the keyword description sequence into a word2Vec pre-training model to perform word vector embedding to obtain a keyword vector;
S302, carrying out similarity calculation on the keyword vector and standard keyword description corresponding to the database, converting the keyword description sequence into a standard keyword sequence through an optimal similarity principle, and carrying out IDF feature extraction on the standard keyword sequence to obtain IDF weight corresponding to the standard keyword sequence;
s303, determining a three-dimensional model based on the corresponding standard keyword description, and selecting corresponding model description semantic features through the three-dimensional model to serve as IDF weights corresponding to the three-dimensional model in a database;
S304, obtaining an inner product corresponding to the three-dimensional model through the IDF weight corresponding to the standard keyword sequence and the IDF weight corresponding to the three-dimensional model in the database, and taking the inner product as the similarity between the output voice and the three-dimensional model in the pre-built database.
7. The method for retrieving a furniture model based on speech and sketch according to claim 6, wherein the calculation of the IDF weight in S304 is as follows:
Where W tag represents the weight, N model represents the number of three-dimensional models, and N tag represents the number of labels corresponding to the standard keyword sequence.
8. The furniture model retrieval method based on voice and sketch according to claim 1, wherein S4 the obtaining the final retrieval result based on the sketch retrieval result and the voice retrieval result through voting comprises:
S401, respectively acquiring the similarity of all three-dimensional models according to the input two-dimensional sketch and the input voice, respectively sequencing all the three-dimensional models from large to small according to the similarity, and distributing weights;
s402, voting the two search results according to the weights to obtain total weights, and sorting the two search results according to the total weights from high to low to obtain a final search result.
CN202210043225.XA 2022-01-14 2022-01-14 Furniture model retrieval method based on voice and sketch Active CN114547357B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210043225.XA CN114547357B (en) 2022-01-14 2022-01-14 Furniture model retrieval method based on voice and sketch

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210043225.XA CN114547357B (en) 2022-01-14 2022-01-14 Furniture model retrieval method based on voice and sketch

Publications (2)

Publication Number Publication Date
CN114547357A CN114547357A (en) 2022-05-27
CN114547357B true CN114547357B (en) 2024-08-16

Family

ID=81672138

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210043225.XA Active CN114547357B (en) 2022-01-14 2022-01-14 Furniture model retrieval method based on voice and sketch

Country Status (1)

Country Link
CN (1) CN114547357B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101004748A (en) * 2006-10-27 2007-07-25 北京航空航天大学 Method for searching 3D model based on 2D sketch
CN102955848A (en) * 2012-10-29 2013-03-06 北京工商大学 Semantic-based three-dimensional model retrieval system and method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20130059040A (en) * 2011-11-28 2013-06-05 연세대학교 산학협력단 Article retrival system and method using sketch in store

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101004748A (en) * 2006-10-27 2007-07-25 北京航空航天大学 Method for searching 3D model based on 2D sketch
CN102955848A (en) * 2012-10-29 2013-03-06 北京工商大学 Semantic-based three-dimensional model retrieval system and method

Also Published As

Publication number Publication date
CN114547357A (en) 2022-05-27

Similar Documents

Publication Publication Date Title
CN109829104B (en) Semantic similarity based pseudo-correlation feedback model information retrieval method and system
CN112905822B (en) Deep supervision cross-modal counterwork learning method based on attention mechanism
CN104376406B (en) A kind of enterprise innovation resource management and analysis method based on big data
CN117290489B (en) Method and system for quickly constructing industry question-answer knowledge base
CN102207945B (en) Knowledge network-based text indexing system and method
CN110188197B (en) Active learning method and device for labeling platform
CN104199965A (en) Semantic information retrieval method
CN111061939B (en) Scientific research academic news keyword matching recommendation method based on deep learning
CN112559684A (en) Keyword extraction and information retrieval method
CN114139533A (en) Text content auditing method for Chinese novel field
CN112307182A (en) Question-answering system-based pseudo-correlation feedback extended query method
CN110990597A (en) Cross-modal data retrieval system based on text semantic mapping and retrieval method thereof
CN115203421A (en) Method, device and equipment for generating label of long text and storage medium
CN118069812B (en) Navigation method based on large model
CN111104437A (en) Test data unified retrieval method and system based on object model
CN118245564B (en) Method and device for constructing feature comparison library supporting semantic review and repayment
CN116450883A (en) Video moment retrieval method based on video content fine granularity information
CN117993393A (en) Method, device and system for checking online labeling policy terms based on word and sentence vectors
CN114493783A (en) Commodity matching method based on double retrieval mechanism
CN117454898A (en) Method and device for realizing legal entity standardized output according to input text
CN105117735A (en) Image detection method in big data environment
Zhang et al. mSHINE: A multiple-meta-paths simultaneous learning framework for heterogeneous information network embedding
CN114547357B (en) Furniture model retrieval method based on voice and sketch
CN114298020B (en) Keyword vectorization method based on topic semantic information and application thereof
CN110688559A (en) Retrieval method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant