CN109033321A - It is a kind of that image is with natural language feature extraction and the language based on keyword indicates image partition method - Google Patents

It is a kind of that image is with natural language feature extraction and the language based on keyword indicates image partition method Download PDF

Info

Publication number
CN109033321A
CN109033321A CN201810790480.4A CN201810790480A CN109033321A CN 109033321 A CN109033321 A CN 109033321A CN 201810790480 A CN201810790480 A CN 201810790480A CN 109033321 A CN109033321 A CN 109033321A
Authority
CN
China
Prior art keywords
feature
image
keyword
word
language
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810790480.4A
Other languages
Chinese (zh)
Other versions
CN109033321B (en
Inventor
李宏亮
石恒璨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Quick Eye Technology Co Ltd
Original Assignee
Chengdu Quick Eye Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Quick Eye Technology Co Ltd filed Critical Chengdu Quick Eye Technology Co Ltd
Priority to CN201810790480.4A priority Critical patent/CN109033321B/en
Publication of CN109033321A publication Critical patent/CN109033321A/en
Application granted granted Critical
Publication of CN109033321B publication Critical patent/CN109033321B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Abstract

The present invention provides a kind of image and natural language feature extraction and the language based on keyword indicates image partition method, on the basis of image characteristics extraction and natural language feature extraction, to the natural language of input picture and input, the keyword according to included in natural language, by the feature f of keyword correspondence image region ii, keyword weighting after sentence feature qiAnd the corresponding vision contextual feature c based on keywordiThree features are cascaded altogether;Feature input multi-layer perception (MLP) after cascade is classified, and segmentation result is obtained.Compared with prior art, the feature extraction of image and natural language can be easy to implement the language instruction image partition method based on keyword;Language of the invention indicates image partition method, reduces the processing difficulty to long sentence, improves the accuracy of object positioning and identification, and then improves language instruction image segmentation precision.

Description

It is a kind of that image is with natural language feature extraction and the language based on keyword indicates image Dividing method
Technical field
The present invention relates to a kind of image and natural language feature extraction and the language based on keyword indicates image segmentation side Method is related to image procossing, computer vision, image segmentation, language and image Combined Treatment and leads.
Background technique
With the arrival of big data era, different types of mass data circulates in network, by different types of data phase In conjunction with being the new demand of big data era.Wherein, image procossing has received widespread attention in conjunction with natural language.Language instruction Image segmentation refers to, is partitioned into the object of natural language description in image, is the committed step in language and image Combined Treatment.
The technology for solving language instruction image segmentation problem at present mainly utilizes deep neural network to extract nature respectively Language and characteristics of image, then by natural language and feature combinations, image is split as new feature.It specifically can be with It is divided into two classes: the language instruction image partition method based on sentence and the language instruction image partition method based on word.It is based on The language instruction image partition method of sentence extracts the feature of entire sentence, with feature combinations;Language based on word refers to Show that image partition method extracts the feature of each word, by its respectively with feature combinations.These methods are primarily present two Defect:
1, the difference of importance between word is had ignored, each word of equality processing causes to be difficult to handle to long sentence;
2, the context relations such as appearance, the position inside image between different zones are not accounted for, and these vision contexts close It is often most important for finding the object of natural language description in the picture.
Summary of the invention
The present invention provides a kind of images and natural language feature extracting method, and having can be easy to implement based on keyword Language indicate image partition method the characteristics of.
The present invention also provides a kind of, and the language based on keyword indicates image partition method, and having reduces to long sentence Processing difficulty improves the characteristics of object is positioned at the accuracy of identification.
A kind of image for there is provided according to the present invention and natural language feature extracting method, including image characteristic extracting method and Natural language feature extracting method;Wherein,
Image characteristic extracting method includes, and for input picture, extracts characteristics of image F using depth convolutional neural networks;Institute Stating characteristics of image is a two dimensional character figure, the inside each feature vector fiEncode the feature of corresponding region i in image;Root The location information of object is needed according to natural language feature instruction image segmentation task;
Natural language feature extracting method includes, for the natural language of input, by each word be encoded to one-hot feature to After amount, dimensionality reduction is carried out with word insertion;Word after dimensionality reduction presses the sequence in former sentence, sequentially inputs Recognition with Recurrent Neural Network;For T-th of word in sentence, Recognition with Recurrent Neural Network learn the feature q to wordt;The feature q of the wordtEncode word t itself Contextual information with entire sentence of semantic information and word t itself;The feature vector of various words constitutes a matrix Q indicates the feature of entire sentence.
The method of the location information that object is needed according to natural language feature instruction image segmentation task includes extracting The relative position coordinates of each image-region cascade with feature F, obtain the final visual signature V of each image-region.
A kind of language instruction image partition method based on keyword provided according to the present invention, based on above-mentioned image and certainly Right language feature extracting method realizes, specific method includes,
To the natural language of input picture and input, the keyword according to included in natural language, by keyword correspondence image The feature f of region ii, keyword weighting after sentence feature qiAnd the corresponding vision contextual feature c based on keywordiAltogether Three features are cascaded;Feature input multi-layer perception (MLP) after cascade is classified, and segmentation result is obtained;
The multi-layer perception (MLP) is made of two layers of neural network, and first layer includes ReLU activation primitive, and the second layer includes sigmoid Activation primitive;
Wherein, the acquisition methods of image-region i corresponding to keyword include, for key corresponding to each image-region i Word is trained extraction, and the process that training is extracted includes,
For the feature of obtained each word, keyword is extracted using language attention model;The language attention model It is made of two layers of neural network, first layer includes tanh activation primitive, and the second layer does not have activation primitive;For each image-region I first cascades the feature of each word t and the image-region, and then input language attention model, the power that gains attention are beaten Point;To the attention marking be normalized, the marking value after normalization between 0 to 1, closer to 1 illustrate word t for Image-region i is more crucial;Conversely, illustrating that word t is more inessential for image-region i closer to 0;
It is given a mark with attention and corrects the feature of sentence, improved influence power of the keyword in sentence, reduce the influence of non-key word Power;Attention after normalization is given a mark and corresponding word feature qtIt is multiplied, word feature is weighted;Then all words are added Feature after power is added, and generates entire sentence for the sentence feature q of image-region ii
It sets a keyword screening threshold value and illustrates image district if the attention marking after normalization is greater than the threshold value Domain i thinks that word t is keyword;
For each word t, all image-regions for thinking that it is keyword are found out, the context relation in these regions is learnt; The feature in these regions is averaged first, for integrating area information;Then with a full articulamentum based on the spy after being averaged Sign study contextual feature gt
Learn vision contextual feature g corresponding to each keywordtAfterwards, be integrated into entire sentence it is corresponding visually under Literary feature;For image-region i, the vision contextual feature g of corresponding each keywordtIt is added, generates entire sentence pair The vision contextual feature c answeredi
The method also includes being normalized attention marking using softmax.
The method also includes it is 0.05 that setting keyword, which screens threshold value,.
Compared with prior art, the feature extraction of image and natural language can be easy to implement the language based on keyword Indicate image partition method;Language of the invention indicates image partition method, reduces the processing difficulty to long sentence, improves The accuracy of object positioning and identification, and then improve language instruction image segmentation precision.
Detailed description of the invention
Fig. 1 is the schematic illustration of a wherein embodiment of the invention.
Specific embodiment
In order to make the objectives, technical solutions, and advantages of the present invention clearer, with reference to the accompanying drawings and embodiments, right The present invention is further elaborated.It should be appreciated that described herein, specific examples are only used to explain the present invention, not For limiting the present invention.
Any feature disclosed in this specification (including abstract and attached drawing) unless specifically stated can be equivalent by other Or the alternative features with similar purpose are replaced.That is, unless specifically stated, each feature is a series of equivalent or class Like an example in feature.
A kind of image and natural language feature extracting method, including image characteristic extracting method and natural language feature extraction Method;Wherein,
Image characteristic extracting method includes, and for input picture, extracts characteristics of image using depth convolutional neural networks (CNN) F;Described image is characterized in a two dimensional character figure, the inside each feature vector fiEncode the spy of corresponding region i in image Sign;Indicate that image segmentation task needs the location information of object according to natural language feature;
Natural language feature extracting method includes, for the natural language of input, by each word be encoded to one-hot feature to After amount, dimensionality reduction is carried out with word insertion (word embedding);Word after dimensionality reduction presses the sequence in former sentence, sequentially inputs and follows Ring neural network (RNN);For t-th of word in sentence, Recognition with Recurrent Neural Network learns the feature q to wordt;The word Feature qtEncode contextual information of semantic information and the word t itself of word t with entire sentence itself;Various words Feature vector constitute a matrix Q, indicate the feature of entire sentence.
Based on above-mentioned image characteristic extracting method and natural language feature extracting method, it is easy to implement the language based on keyword Speech instruction image partition method.
It is described to indicate that image segmentation task needs object according to natural language feature as one embodiment of the present invention The method of location information include extracting the relative position coordinates of each image-region, cascaded with feature F, obtain each image The final visual signature V in region.
As shown in Figure 1, a kind of language based on keyword indicates image partition method, it is based on above-mentioned image and natural language Feature extracting method realizes that specific method includes,
To the natural language of input picture and input, the keyword according to included in natural language, by keyword correspondence image The feature f of region ii, keyword weighting after sentence feature qiAnd the corresponding vision contextual feature c based on keywordiAltogether Three features are cascaded;Feature input multi-layer perception (MLP) (MLP) after cascade is classified, and segmentation result is obtained;
The multi-layer perception (MLP) is made of two layers of neural network, and first layer includes ReLU activation primitive, and the second layer includes sigmoid Activation primitive;
Wherein, the acquisition methods of image-region i corresponding to keyword include, for key corresponding to each image-region i Word is trained extraction, and the process that training is extracted includes,
For the feature of obtained each word, keyword is extracted using language attention model;The language attention model It is made of two layers of neural network, first layer includes tanh activation primitive, and the second layer does not have activation primitive;For each image-region I first cascades the feature of each word t and the image-region, and then input language attention model, the power that gains attention are beaten Point;To the attention marking be normalized, the marking value after normalization between 0 to 1, closer to 1 illustrate word t for Image-region i is more crucial;Conversely, illustrating that word t is more inessential for image-region i closer to 0;
It is given a mark with attention and corrects the feature of sentence, improved influence power of the keyword in sentence, reduce the influence of non-key word Power;Attention after normalization is given a mark and corresponding word feature qtIt is multiplied, word feature is weighted;Then all words are added Feature after power is added, and generates entire sentence for the sentence feature q of image-region ii
It sets a keyword screening threshold value and illustrates image district if the attention marking after normalization is greater than the threshold value Domain i thinks that word t is keyword;
For each word t, all image-regions for thinking that it is keyword are found out, the context relation in these regions is learnt; The feature in these regions is averaged first, for integrating area information;Then with a full articulamentum based on the spy after being averaged Sign study contextual feature gt
Learn vision contextual feature g corresponding to each keywordtAfterwards, be integrated into entire sentence it is corresponding visually under Literary feature;For image-region i, the vision contextual feature g of corresponding each keywordtIt is added, generates entire sentence pair The vision contextual feature c answeredi
Prior art one side equality handles each of sentence word, causes to be difficult to handle to long sentence;Another party Face does not account for the context relations such as appearance, the position inside image between different zones, these context relations are for scheming It is positioned as in and the object of identification natural language description is most important.The present invention proposes the language instruction image point based on keyword Algorithm is cut, the keyword in natural language is extracted, to reduce the processing difficulty to long sentence.And learn based on keyword Vision context relation improves the accuracy of object positioning and identification, and then improves language and indicate image segmentation precision.
As one embodiment of the present invention, the method also includes returning attention marking using softmax One change processing.
As one embodiment of the present invention, the method also includes it is 0.05 that setting keyword, which screens threshold value,.
It is described in more detail below with a specific embodiment.
It determines database, determines that language indicates image segmentation database, such as Google Referit database.
Data prediction pre-processes database, extracts original image, natural language, segmentation result.It is wherein original Image need to extract the relative position coordinates of each point.Natural language need to convert one-hot vector for word each in sentence.
Build depth network model.Wherein convolutional neural networks (CNN) select DeepLab101, export 60 × 60 figures As region, each provincial characteristics fi is set as 1000 dimensions.Recognition with Recurrent Neural Network (RNN) selects long memory unit (LSTM) in short-term, Every sentence maximum word number is set as 20, and each word feature qt is set as 1000 dimensions.
Determine keyword threshold value.Keyword threshold value Thr is set as 0.05.
Model initialization, the model initialization parameter of convolutional neural networks (CNN) pre-training on ImageNet.Model its Remaining part divides random initializtion.
Learning rate is set and gradient decline strategy, convolutional neural networks (CNN), are based on keyword at language attention model Vision context relational learning model, multi-layer perception (MLP) (MLP) learning rate be set as 0.0001, Recognition with Recurrent Neural Network (RNN) learning rate is set as 0.001.Optimal way is using ADAM gradient decline strategy.
Training pattern, model buildings and initialization finish, and after determining that study and gradient decline are tactful, are trained.It will count Model, 5 epoch of training are sequentially input according to the data of training set in library.
Test model, after model training, the image and sentence of test set in input database obtain language instruction figure As segmentation result.

Claims (5)

1. a kind of image and natural language feature extracting method, including image characteristic extracting method and natural language feature extraction side Method;Wherein,
Image characteristic extracting method includes, and for input picture, extracts characteristics of image F using depth convolutional neural networks;Institute Stating characteristics of image is a two dimensional character figure, the inside each feature vector fiEncode the feature of corresponding region i in image;Root The location information of object is needed according to natural language feature instruction image segmentation task;
Natural language feature extracting method includes, for the natural language of input, by each word be encoded to one-hot feature to After amount, dimensionality reduction is carried out with word insertion;Word after dimensionality reduction presses the sequence in former sentence, sequentially inputs Recognition with Recurrent Neural Network;For T-th of word in sentence, Recognition with Recurrent Neural Network learn the feature q to wordt;The feature q of the wordtEncode word t itself Contextual information with entire sentence of semantic information and word t itself;The feature vector of various words constitutes a matrix Q indicates the feature of entire sentence.
2. image according to claim 1 and natural language feature extracting method, described to be indicated according to natural language feature It includes extracting the relative position coordinates of each image-region that image segmentation task, which needs the method for the location information of object, with spy F cascade is levied, the final visual signature V of each image-region is obtained.
3. a kind of language based on keyword indicates image partition method, based on image of any of claims 1 or 2 and natural language Say that feature extracting method realizes that specific method includes,
To the natural language of input picture and input, the keyword according to included in natural language, by keyword correspondence image The feature f of region ii, keyword weighting after sentence feature qiAnd the corresponding vision contextual feature c based on keywordiAltogether Three features are cascaded;Feature input multi-layer perception (MLP) after cascade is classified, and segmentation result is obtained;
The multi-layer perception (MLP) is made of two layers of neural network, and first layer includes ReLU activation primitive, and the second layer includes sigmoid Activation primitive;
Wherein, the acquisition methods of image-region i corresponding to keyword include, for key corresponding to each image-region i Word is trained extraction, and the process that training is extracted includes,
For the feature of obtained each word, keyword is extracted using language attention model;The language attention model It is made of two layers of neural network, first layer includes tanh activation primitive, and the second layer does not have activation primitive;For each image-region I first cascades the feature of each word t and the image-region, and then input language attention model, the power that gains attention are beaten Point;To the attention marking be normalized, the marking value after normalization between 0 to 1, closer to 1 illustrate word t for Image-region i is more crucial;Conversely, illustrating that word t is more inessential for image-region i closer to 0;
It is given a mark with attention and corrects the feature of sentence, improved influence power of the keyword in sentence, reduce the influence of non-key word Power;Attention after normalization is given a mark and corresponding word feature qtIt is multiplied, word feature is weighted;Then all words are added Feature after power is added, and generates entire sentence for the sentence feature q of image-region ii
It sets a keyword screening threshold value and illustrates image district if the attention marking after normalization is greater than the threshold value Domain i thinks that word t is keyword;
For each word t, all image-regions for thinking that it is keyword are found out, the context relation in these regions is learnt; The feature in these regions is averaged first, for integrating area information;Then with a full articulamentum based on the spy after being averaged Sign study contextual feature gt
Learn vision contextual feature g corresponding to each keywordtAfterwards, be integrated into entire sentence it is corresponding visually under Literary feature;For image-region i, the vision contextual feature g of corresponding each keywordtIt is added, generates entire sentence pair The vision contextual feature c answeredi
4. language according to claim 3 indicates image partition method, the method also includes using softmax to note Meaning power marking is normalized.
5. language according to claim 3 or 4 indicates image partition method, the method also includes setting keyword is sieved Selecting threshold value is 0.05.
CN201810790480.4A 2018-07-18 2018-07-18 Image and natural language feature extraction and keyword-based language indication image segmentation method Active CN109033321B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810790480.4A CN109033321B (en) 2018-07-18 2018-07-18 Image and natural language feature extraction and keyword-based language indication image segmentation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810790480.4A CN109033321B (en) 2018-07-18 2018-07-18 Image and natural language feature extraction and keyword-based language indication image segmentation method

Publications (2)

Publication Number Publication Date
CN109033321A true CN109033321A (en) 2018-12-18
CN109033321B CN109033321B (en) 2021-12-17

Family

ID=64643921

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810790480.4A Active CN109033321B (en) 2018-07-18 2018-07-18 Image and natural language feature extraction and keyword-based language indication image segmentation method

Country Status (1)

Country Link
CN (1) CN109033321B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109711463A (en) * 2018-12-25 2019-05-03 广东顺德西安交通大学研究院 Important object detection method based on attention
CN111598155A (en) * 2020-05-13 2020-08-28 北京工业大学 Fine-grained image weak supervision target positioning method based on deep learning
CN112037239A (en) * 2020-08-28 2020-12-04 大连理工大学 Text guidance image segmentation method based on multi-level explicit relation selection
CN114299348A (en) * 2022-02-21 2022-04-08 山东力聚机器人科技股份有限公司 Image classification method and device based on restoration self-supervision task

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103559193A (en) * 2013-09-10 2014-02-05 浙江大学 Topic modeling method based on selected cell
CN106227851A (en) * 2016-07-29 2016-12-14 汤平 Based on the image search method searched for by depth of seam division that degree of depth convolutional neural networks is end-to-end
CN106599198A (en) * 2016-12-14 2017-04-26 广东顺德中山大学卡内基梅隆大学国际联合研究院 Image description method for multi-stage connection recurrent neural network
CN106778835A (en) * 2016-11-29 2017-05-31 武汉大学 The airport target by using remote sensing image recognition methods of fusion scene information and depth characteristic
CN107391709A (en) * 2017-07-28 2017-11-24 深圳市唯特视科技有限公司 A kind of method that image captions generation is carried out based on new attention model
CN107608943A (en) * 2017-09-08 2018-01-19 中国石油大学(华东) Merge visual attention and the image method for generating captions and system of semantic notice
CN107688821A (en) * 2017-07-11 2018-02-13 西安电子科技大学 View-based access control model conspicuousness and across the modality images natural language description methods of semantic attribute
US9939272B1 (en) * 2017-01-06 2018-04-10 TCL Research America Inc. Method and system for building personalized knowledge base of semantic image segmentation via a selective random field approach
CN107909115A (en) * 2017-12-04 2018-04-13 上海师范大学 A kind of image Chinese subtitle generation method
CN108009154A (en) * 2017-12-20 2018-05-08 哈尔滨理工大学 A kind of image Chinese description method based on deep learning model
CN108228686A (en) * 2017-06-15 2018-06-29 北京市商汤科技开发有限公司 It is used to implement the matched method, apparatus of picture and text and electronic equipment

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103559193A (en) * 2013-09-10 2014-02-05 浙江大学 Topic modeling method based on selected cell
CN106227851A (en) * 2016-07-29 2016-12-14 汤平 Based on the image search method searched for by depth of seam division that degree of depth convolutional neural networks is end-to-end
CN106778835A (en) * 2016-11-29 2017-05-31 武汉大学 The airport target by using remote sensing image recognition methods of fusion scene information and depth characteristic
CN106599198A (en) * 2016-12-14 2017-04-26 广东顺德中山大学卡内基梅隆大学国际联合研究院 Image description method for multi-stage connection recurrent neural network
US9939272B1 (en) * 2017-01-06 2018-04-10 TCL Research America Inc. Method and system for building personalized knowledge base of semantic image segmentation via a selective random field approach
CN108228686A (en) * 2017-06-15 2018-06-29 北京市商汤科技开发有限公司 It is used to implement the matched method, apparatus of picture and text and electronic equipment
CN107688821A (en) * 2017-07-11 2018-02-13 西安电子科技大学 View-based access control model conspicuousness and across the modality images natural language description methods of semantic attribute
CN107391709A (en) * 2017-07-28 2017-11-24 深圳市唯特视科技有限公司 A kind of method that image captions generation is carried out based on new attention model
CN107608943A (en) * 2017-09-08 2018-01-19 中国石油大学(华东) Merge visual attention and the image method for generating captions and system of semantic notice
CN107909115A (en) * 2017-12-04 2018-04-13 上海师范大学 A kind of image Chinese subtitle generation method
CN108009154A (en) * 2017-12-20 2018-05-08 哈尔滨理工大学 A kind of image Chinese description method based on deep learning model

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
LIANG-CHIEH CHEN ET AL.: "Attention to Scale: Scale-aware Semantic Image Segmentation", 《2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
RONGHANG HU ET AL.: "Segmentation from Natural Language", 《EUROPEAN CONFERENCE ON COMPUTER VISION》 *
李志欣 等: "图像检索中语义映射方法综述", 《计算机辅助设计与图形学学报》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109711463A (en) * 2018-12-25 2019-05-03 广东顺德西安交通大学研究院 Important object detection method based on attention
CN109711463B (en) * 2018-12-25 2023-04-07 广东顺德西安交通大学研究院 Attention-based important object detection method
CN111598155A (en) * 2020-05-13 2020-08-28 北京工业大学 Fine-grained image weak supervision target positioning method based on deep learning
CN112037239A (en) * 2020-08-28 2020-12-04 大连理工大学 Text guidance image segmentation method based on multi-level explicit relation selection
CN114299348A (en) * 2022-02-21 2022-04-08 山东力聚机器人科技股份有限公司 Image classification method and device based on restoration self-supervision task

Also Published As

Publication number Publication date
CN109033321B (en) 2021-12-17

Similar Documents

Publication Publication Date Title
CN110348319B (en) Face anti-counterfeiting method based on face depth information and edge image fusion
CN109033321A (en) It is a kind of that image is with natural language feature extraction and the language based on keyword indicates image partition method
CN111177446B (en) Method for searching footprint image
CN111126069B (en) Social media short text named entity identification method based on visual object guidance
CN107273864B (en) Face detection method based on deep learning
CN106960206A (en) Character identifying method and character recognition system
RU2707147C1 (en) Neural network training by means of specialized loss functions
CN112733866A (en) Network construction method for improving text description correctness of controllable image
CN107491729B (en) Handwritten digit recognition method based on cosine similarity activated convolutional neural network
CN110781897A (en) Semantic edge detection method based on deep learning
CN112560710B (en) Method for constructing finger vein recognition system and finger vein recognition system
CN109885796A (en) A kind of Internet news figure matching detection method based on deep learning
Yu et al. Exemplar-based recursive instance segmentation with application to plant image analysis
CN111126155B (en) Pedestrian re-identification method for generating countermeasure network based on semantic constraint
CN109508640A (en) A kind of crowd's sentiment analysis method, apparatus and storage medium
CN117149944A (en) Multi-mode situation emotion recognition method and system based on wide time range
CN115170403A (en) Font repairing method and system based on deep meta learning and generation countermeasure network
CN110503090A (en) Character machining network training method, character detection method and character machining device based on limited attention model
CN113449776A (en) Chinese herbal medicine identification method and device based on deep learning and storage medium
CN116385832A (en) Bimodal biological feature recognition network model training method
EP4288910A1 (en) Continual learning neural network system training for classification type tasks
WO2020224244A1 (en) Method and apparatus for obtaining depth-of-field image
CN113569867A (en) Image processing method and device, computer equipment and storage medium
CN113129399A (en) Pattern generation
CN113792703B (en) Image question-answering method and device based on Co-Attention depth modular network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant