CN117094966A - Tongue image identification method and device based on image amplification and computer equipment - Google Patents

Tongue image identification method and device based on image amplification and computer equipment Download PDF

Info

Publication number
CN117094966A
CN117094966A CN202311051530.4A CN202311051530A CN117094966A CN 117094966 A CN117094966 A CN 117094966A CN 202311051530 A CN202311051530 A CN 202311051530A CN 117094966 A CN117094966 A CN 117094966A
Authority
CN
China
Prior art keywords
tongue
image
model
images
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311051530.4A
Other languages
Chinese (zh)
Other versions
CN117094966B (en
Inventor
冯健
陈栋栋
赖永航
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qingdao Medcare Digital Engineering Co ltd
Original Assignee
Qingdao Medcare Digital Engineering Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qingdao Medcare Digital Engineering Co ltd filed Critical Qingdao Medcare Digital Engineering Co ltd
Priority to CN202311051530.4A priority Critical patent/CN117094966B/en
Publication of CN117094966A publication Critical patent/CN117094966A/en
Application granted granted Critical
Publication of CN117094966B publication Critical patent/CN117094966B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/778Active pattern-learning, e.g. online learning of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Abstract

The invention relates to the technical field of image processing, and provides a tongue image recognition method, a tongue image recognition device and computer equipment based on image amplification, wherein the method comprises the following steps: carrying out effective region segmentation processing on an initial tongue image to be processed to obtain a first tongue image and a first tongue outline mask image; generating N first generated tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, wherein N is greater than or equal to 1; respectively carrying out color correction on the first tongue images based on the N first generation type tongue images, and respectively generating N second tongue images corresponding to the N first generation type tongue images one by one; and taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image. The tongue image recognition method and device improve the accuracy of tongue image recognition.

Description

Tongue image identification method and device based on image amplification and computer equipment
Technical Field
The present invention relates to the field of image processing technologies, and in particular, to a tongue image recognition method, apparatus and computer device based on image amplification.
Background
The tongue is one of the important constituent organs in the human digestive tract, consists of a plurality of crisscrossed striated muscles, and is externally covered with a special mucous membrane structure. In traditional Chinese medicine, the tongue is closely related to the viscera, especially the spleen and stomach, so that the tongue is often used to infer the digestive tract diseases.
When digestive tract diseases occur, the characteristics of tongue, coating, tooth trace and the like often show regular changes, and the changes can be captured and analyzed by naked eyes or by an image. In modern medicine, the doctor is assisted in identifying digestive tract diseases by judging the type of digestive tract diseases corresponding to tongue images. However, in the remote consultation, because the tongue coating image shot by the mobile phone is easily affected by factors such as a mobile phone imaging algorithm, light, background interference and the like, the image cannot reflect real image information, such as white tongue coating, but the tongue coating shot by the mobile phone is reddish, and the error of tongue image recognition is caused by the deviation caused by shooting.
Disclosure of Invention
The present invention has been made in view of the above problems, and has as its object to provide a tongue image recognition method, apparatus and computer device based on image augmentation, which overcome the above problems.
In one aspect of the present invention, there is provided a tongue image recognition method based on image augmentation, the method comprising:
carrying out effective region segmentation processing on an initial tongue image to be processed to obtain a first tongue image and a first tongue outline mask image;
generating N first generated tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, wherein N is greater than or equal to 1;
respectively carrying out color correction on the first tongue images based on N first generation type tongue images, and respectively generating N second tongue images corresponding to the N first generation type tongue images one by one;
and taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image.
Further, the performing color correction on the first tongue images based on the N first generated tongue images, and generating N second tongue images corresponding to the N first generated tongue images one to one respectively includes:
sequentially selecting one image from the first generated tongue image as a target generated tongue image, and performing feature point matching on the first tongue image and the target generated tongue image to obtain feature point pairs;
Generating a color correction matrix between the first tongue image and the target generated tongue image according to the color corresponding relation of the feature points;
and performing color correction on the first tongue image based on the color correction matrix to obtain a second tongue image corresponding to the target generated tongue image.
Further, the image recognition model is a cascading network model, a first network model of the cascading network model is a text description prediction model, and a second network model of the cascading network model is a tongue image classification prediction model;
the step of calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image comprises the following steps:
inputting each piece of target image data into the text description prediction model for recognition so as to obtain text descriptions corresponding to each piece of target image data, wherein the text descriptions are used for describing tongue texture features in the target image data;
and inputting text descriptions corresponding to the target image data into the tongue image classification prediction model for recognition so as to obtain a recognition result corresponding to the initial tongue image.
Further, before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further includes: pre-training the text description prediction model;
The pre-training the text description prediction model specifically comprises the following steps:
carrying out effective area segmentation processing on the first initial tongue image sample to obtain a first tongue image sample;
acquiring a text description describing tongue texture features of the first tongue image sample;
and inputting the first tongue image sample and the text description of the first tongue image sample as training data into an input layer of a preset multi-mode training model, and obtaining a text description prediction model through training.
Further, before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further includes: pre-training the tongue image classification prediction model;
the pre-training the tongue image classification prediction model specifically comprises the following steps:
carrying out effective region segmentation processing on a second initial tongue image sample to obtain a second tongue image sample and a first tongue outline mask image sample, and carrying out tongue image classification labeling on the second initial tongue image sample;
generating N first generation type tongue image samples corresponding to the second tongue outline mask image based on a preset image diffusion model;
Respectively carrying out color correction on the second tongue image samples based on N first generation type tongue image samples, and respectively generating N third tongue image samples which are in one-to-one correspondence with the N first generation type tongue image samples;
taking N third tongue image samples or a set of N third tongue image samples and second tongue image samples as target image data samples, and calling a preset text description prediction model to identify the target image data samples so as to obtain text descriptions corresponding to the target image data samples;
and inputting a tongue image classification result of the second initial tongue image sample and text descriptions corresponding to each target image data sample as training data into an input layer of a preset tongue image classification prediction training model, and obtaining tongue image classification through training.
Further, before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further includes: pre-training the tongue image classification prediction model;
the pre-training the tongue image classification prediction model specifically comprises the following steps:
acquiring text descriptions of a plurality of tongue texture features corresponding to different tongue image classification results;
And inputting any target tongue image classification result and text description of a plurality of tongue texture features corresponding to the current target tongue image classification result as training data into an input layer of a preset tongue image classification prediction training model, and obtaining the tongue image classification prediction model through training.
Further, the image recognition model is an image recognition model obtained by training in advance, and the calling the preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image comprises the following steps:
inputting each target image data into the image recognition model for recognition so as to obtain tongue image classification results corresponding to each target image data;
and carrying out statistical calculation on tongue image classification results corresponding to each target image data so as to finally obtain an initial tongue image recognition result.
Further, before the effective region segmentation processing is performed on the initial tongue image to be processed to obtain the first tongue image and the first tongue contour mask image, the method further includes: pre-training a tongue segmentation model;
the pre-trained tongue segmentation model comprises:
acquiring a third initial tongue image sample containing a tongue effective area;
Labeling the tongue effective areas in each third initial tongue image sample to obtain a semantic segmentation training sample;
inputting the semantic segmentation training sample into an input layer of a preset semantic segmentation model, and obtaining a tongue segmentation model through training.
Further, before the effective region segmentation processing is performed on the initial tongue image to be processed to obtain a first tongue image and a first tongue outline mask image, the method further comprises;
judging whether the initial tongue image contains a tongue effective area, and if not, sending out an error prompt.
In another aspect of the present invention, there is provided a tongue image recognition apparatus based on image augmentation, the apparatus comprising:
the effective area segmentation module is used for carrying out effective area segmentation processing on the initial tongue image to be processed to obtain a first tongue image and a first tongue outline mask image;
the image diffusion module is used for generating N first generated tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, wherein N is greater than or equal to 1;
the color correction module is used for respectively carrying out color correction on the first tongue images based on N first generation type tongue images and respectively generating N second tongue images which are in one-to-one correspondence with the N first generation type tongue images;
The image recognition module is used for taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image.
In another aspect of the application, a computer device is provided comprising a memory, a processor, and a computer program stored on the memory and executable on the processor; the computer program, when executed by the processor, implements the steps of the tongue image recognition method based on image augmentation as described in any one of the above.
After an initial tongue image is obtained, effective area segmentation processing is carried out on the initial tongue image to obtain a first tongue image and a first tongue outline mask image; generating N first generation type tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, respectively carrying out color correction on the first tongue images based on the N first generation type tongue images, respectively generating N second tongue images corresponding to the N first generation type tongue images one by one, and improving generalization capability and robustness in image recognition by acquiring the N second tongue images after color correction; and finally, taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image.
The foregoing description is only an overview of the present invention, and is intended to be implemented in accordance with the teachings of the present invention in order that the same may be more clearly understood and to make the same and other objects, features and advantages of the present invention more readily apparent.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the preferred embodiments. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the invention. In the drawings:
FIG. 1 is a flowchart of a tongue image recognition method based on image augmentation according to an embodiment of the present invention;
FIG. 2 is a schematic illustration of a tongue effective area image and a tongue contour mask of a tongue image according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a tongue image recognition device based on image augmentation according to an embodiment of the present invention.
Detailed Description
Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
It will be understood by those skilled in the art that all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs unless defined otherwise. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the prior art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
Example 1
The embodiment of the invention provides a tongue image recognition method based on image amplification, which is shown in fig. 1 and comprises the following steps:
s1, carrying out effective area segmentation processing on an initial tongue image to be processed to obtain a first tongue image and a first tongue outline mask image;
specifically, the outer contour of the tongue region can be used as an image boundary to carry out effective region segmentation, the region in the contour is used for obtaining an effective tongue image, the region outside the contour is filled with black pixels to generate a new image, the new image is cut off an ineffective region to obtain a tongue contour mask image, and part of effective characteristic information of the tongue is still reserved in the tongue contour mask image.
S2, generating N first generated tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, wherein N is greater than or equal to 1;
the image diffusion technology enables a model to learn the characteristics and the patterns of an image by training a large number of image data sets, and then generates a new generated image, wherein the generated image has the characteristics and the patterns of the learned image. The image diffusion model preset in the invention learns a large number of tongue images capable of truly reflecting tongue fur texture features, so that a first generated tongue image corresponding to a first tongue outline mask image can be generated according to the first tongue outline mask image. The first generated tongue image can truly reflect the texture characteristics of the tongue fur, and has the same structural characteristics as the first tongue image, so that the first tongue image can be used as the basis for color correction.
S3, respectively carrying out color correction on the first tongue images based on N first generation type tongue images, and respectively generating N second tongue images corresponding to the N first generation type tongue images one by one;
s4, taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image.
According to the invention, the first tongue image and the first tongue outline mask image are obtained by carrying out effective area segmentation processing on the initial tongue image, so that noise interference of images outside the tongue effective area in image recognition is reduced, a plurality of first generated tongue images capable of truly reflecting the texture characteristics of the tongue are generated through the image diffusion model and the first tongue outline mask image, and as the first generated tongue image and the first tongue image are in the same position relationship, the first tongue image is subjected to color correction through the first generated tongue image, a plurality of second tongue images subjected to color correction can be obtained, and the second tongue image can truly reflect the texture characteristics of the tongue. Finally, learning and recognition are carried out based on a preset image recognition model, and finally, a recognition result of the tongue fur image picture is obtained, so that the accuracy of image recognition is improved, and misdiagnosis caused by deviation of mobile phone photography is avoided.
In an embodiment of the present invention, after obtaining the identification result, the method further includes: comparing the identification result with a pre-stored initial result, and judging whether the treatment effect is effective or not according to the comparison result; if no effect exists, pushing the initial tongue image and/or the identification result to a preset doctor system. The tongue image recognition method based on image amplification can assist doctors to finish follow-up tracking of the illness state of digestive diseases.
Specifically, the initial tongue image uploaded by the user through the mobile phone is subjected to image diffusion and color correction to obtain a plurality of tongue images after color correction, and the tongue images after color correction and the initial tongue image obtained in the original are input into the image recognition model together, so that the robustness and generalization capability of model prediction are improved, the tongue image of a patient can be accurately recognized, and a doctor can conveniently adjust a treatment scheme according to the treatment effect by checking the treatment effect of the patient at regular time.
In addition, in the embodiment of the invention, the patient can shoot and upload tongue images through the patient condition follow-up APP at the mobile phone end, and the APP can also remind the user to shoot tongue images by the mobile phone at regular time and punch cards during the medication period of the patient.
In step S1, the initial tongue image is subjected to effective region segmentation processing by using a tongue segmentation model obtained by pre-training, the tongue segmentation model uses the outer contour of the tongue region as an image boundary, the outer region of the contour is filled with black pixels to generate a new image, the new image cuts off an ineffective region, and effective characteristic information of the tongue is reserved, so that noise interference in subsequent image recognition can be reduced, and recognition accuracy is improved.
Further, before the effective region segmentation processing is performed on the initial tongue image to obtain the first tongue image and the first tongue outline mask image, the method further includes: pre-training a tongue segmentation model; the pre-trained tongue segmentation model comprises: acquiring a third initial tongue image sample containing a tongue effective area; labeling the tongue effective areas in each third initial tongue image sample to obtain a semantic segmentation training sample; inputting the semantic segmentation training sample into an input layer of a preset semantic segmentation model, and obtaining a tongue segmentation model through training.
In addition, the tongue segmentation model of the embodiment of the present invention may be further used to determine whether a tongue effective area exists in the initial tongue image, specifically, before the effective area segmentation process is performed on the initial tongue image to obtain a first tongue image and a first tongue contour mask image, the method further includes; judging whether the initial tongue image contains a tongue effective area, and if not, sending out an error prompt.
In step S2, the image diffusion model may be a Controlnet image diffusion model, which is trained based on a large number of tongue images capable of truly reflecting tongue fur texture features, and thus may be considered as a tongue image in which the first generated tongue image generated based on the image diffusion model has tongue fur texture features capable of truly reflecting tongue fur.
Further, the embodiment of the invention further includes a training method for an image diffusion model, and before generating at least one first generated tongue image corresponding to the first tongue outline mask image based on a preset image diffusion model, the method further includes: acquiring a fourth initial tongue image sample capable of truly reflecting tongue texture features; carrying out effective region segmentation processing on the fourth initial tongue image sample to obtain a fourth tongue image sample and a second tongue outline mask image sample; and inputting the fourth tongue image sample and the second tongue outline mask image sample as training data into an input layer of a preset image diffusion training model, and obtaining the image diffusion model through training.
Fig. 2 schematically illustrates a training pair of a fourth initial tongue image sample and a second tongue outline mask image according to an embodiment of the present invention, and because the tongue image in the fourth initial tongue image sample can truly reverse the texture features of the tongue, the generated tongue images generated by the trained image diffusion model in the embodiment of the present invention are tongue effective area pictures that can truly reflect the texture features of the tongue. It can be inferred that the first generated tongue images in the embodiments of the present invention are tongue effective area images that can truly reflect the tongue texture features, and the first generated tongue images can be used as references for performing color correction on the first tongue images. So as to carry out color correction on the first tongue image by referencing the color information of the first generated tongue image. The tongue texture features can be texture features of tongue quality and tongue fur.
In step S3, performing color correction on the first tongue images based on the N first generated tongue images, and generating N second tongue images corresponding to the N first generated tongue images one by one includes: sequentially selecting one image from the first generated tongue image as a target generated tongue image, and performing feature point matching on the first tongue image and the target generated tongue image to obtain feature point pairs; generating a color correction matrix between the first tongue image and the target generated tongue image according to the color corresponding relation of the feature points; and performing color correction on the first tongue image based on the color correction matrix to obtain a second tongue image corresponding to the target generated tongue image. That is, the number of second tongue images in the embodiment of the present invention is the same as the number of generated tongue images.
The color correction matrix is a technology for correcting the color deviation of the digital image, and the color correction matrix is more true and accurate by adjusting the color of the image. The basic principle is to build a matrix by measuring RGB values of a series of standard colors and RGB values of the same series of standard colors photographed by a camera, the matrix being capable of correcting color deviations of an image photographed by the camera. The matrix contains color deviation information of an image shot by the camera, and the color of each pixel point of the image can be corrected in a matrix operation mode, so that the color of the image is more real and accurate.
Further, the first generated tongue image in the embodiment of the invention has RGB values capable of truly inverting the tongue texture feature, the SIFT matching algorithm can be selected to perform feature point matching on the target generated tongue image and the first tongue image, the feature point pair is determined as an extraction key matching point, and the RGB values capable of truly reflecting the tongue texture feature are provided in the target generated tongue image, and based on certain consistency of the RGB values of the same object in different images, the color correction matrix can be calculated by matching the RGB values of the same object in different images, and the color correction can be completed by processing the first effective area image by using the color correction matrix.
According to the color correction method based on feature point matching, feature matching between two images with larger difference can be achieved, robustness and accuracy of color correction are greatly improved, and corrected images can show true colors of tongue fur.
In one embodiment of the present invention, the image recognition model in step S4 may be a cascading network model, a first network model of the cascading network model is a text description prediction model, and a second network model of the cascading network model is a tongue image classification model; the step of calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image comprises the following steps: inputting each piece of target image data into the text description prediction model for recognition so as to obtain text descriptions corresponding to each piece of target image data, wherein the text descriptions are used for describing tongue texture features in the target image data; and inputting text descriptions corresponding to the target image data into the tongue image classification model for recognition so as to obtain a recognition result corresponding to the initial tongue image.
In the embodiment of the invention, all second tongue images are used as target image data, wherein all second tongue images are tongue images after color correction. Or the second tongue image and the first tongue image are used as target image data, the original tongue image shot by the mobile phone is used as a tongue image to be learned, if the original image has no chromatic aberration, the texture characteristics of the tongue and tongue fur can be truly represented, and the generalization capability of the model can be further improved by adding the image into the model. The tongue image classification in the embodiment of the invention can classify intestinal diseases of different grades aiming at tongue image characterization, namely, based on the recognition result of the tongue image, doctors can be assisted to know the illness state of patients. It can be appreciated that in the embodiment of the present invention, tongue images may be classified according to other classification criteria, so as to accomplish different types of tongue image recognition.
Further, the embodiment of the invention also comprises a training method for the text description prediction model and the tongue image classification model. Therefore, before calling a preset image recognition model to recognize the target image data so as to obtain the recognition result of the initial tongue image, the method further comprises the following steps: pre-training the text description prediction model; the pre-training the text description prediction model specifically comprises the following steps: carrying out effective area segmentation processing on the first initial tongue image sample to obtain a first tongue image sample; acquiring a text description describing tongue texture features of the first tongue image sample; and inputting the first tongue image sample and the text description of the first tongue image sample as training data into an input layer of a preset multi-mode training model, and obtaining a text description prediction model through training. The text description of the tongue texture feature of the embodiments of the present invention can be expressed as: the tongue is pale white, the tongue shape is old and tender, the tongue coating is thick and dry, the tongue coating is white and the tongue coating is partially concentrated, and the text description of the tongue texture features is only an alternative embodiment of the invention and is not used as the basis for limiting the protection scope of the invention.
Further, before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further includes: pre-training the tongue image classification model; the pre-training the tongue image classification model specifically comprises the following steps: carrying out effective region segmentation processing on a second initial tongue image sample to obtain a second tongue image sample and a first tongue outline mask image sample, and carrying out tongue image classification labeling on the second initial tongue image sample; generating N first generation type tongue image samples corresponding to the second tongue outline mask image based on a preset image diffusion model; respectively carrying out color correction on the second tongue image samples based on N first generation type tongue image samples, and respectively generating N third tongue image samples which are in one-to-one correspondence with the N first generation type tongue image samples; taking N third tongue image samples or a set of N third tongue image samples and second tongue image samples as target image data samples, and calling a preset text description prediction model to identify the target image data samples so as to obtain text descriptions corresponding to the target image data samples; and inputting a tongue image classification result of the second initial tongue image sample and text descriptions corresponding to each target image data sample as training data into an input layer of a preset tongue image classification prediction training model, and obtaining a tongue image classification model through training. The tongue image classification result of the embodiment of the present invention may specifically be a tongue image belonging to a class C1 of chronic atrophic gastritis, a tongue image belonging to a class C2 of chronic atrophic gastritis, or the like.
Further, before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further includes: pre-training the tongue image classification model; the pre-training the tongue image classification model specifically comprises the following steps: acquiring text descriptions of a plurality of tongue texture features corresponding to different tongue image classification results; and inputting any target tongue image classification result and text description of a plurality of tongue texture features corresponding to the current target tongue image classification result as training data into an input layer of a preset tongue image classification prediction training model, and obtaining a tongue image classification model through training.
In one embodiment of the present invention, the image recognition model in step S4 may be an image recognition model, where the image recognition model is an image recognition model obtained by training in advance, and the calling a preset image recognition model to recognize the target image data, so as to obtain a recognition result of the initial tongue image, includes: inputting each target image data into the image recognition model for recognition so as to obtain tongue image classification results corresponding to each target image data; and carrying out statistical calculation on tongue image classification results corresponding to each target image data so as to finally obtain an initial tongue image recognition result. According to the embodiment, the target image data are directly subjected to learning identification based on the image identification model, after tongue image classification results corresponding to each target image data are obtained, the tongue image classification results with the largest occurrence number are counted and selected as the identification results of finally identified tongue fur image pictures.
For the purposes of simplicity of explanation, the methodologies are shown and described as a series of acts, it is to be understood and appreciated by one of ordinary skill in the art that the methodologies are not limited by the order of acts, as some acts may, in accordance with the methodologies, take place in other order or concurrently. Further, those skilled in the art will appreciate that the embodiments described in the specification are presently preferred embodiments, and that the acts are not necessarily required by the embodiments of the invention.
Example two
Fig. 3 schematically illustrates a schematic structural diagram of a tongue image recognition device based on image augmentation, and referring to fig. 3, the tongue image recognition device based on image augmentation in the embodiment of the present invention specifically includes an effective area segmentation module 301, an image diffusion module 302, a color correction module 303, and an image recognition module 304, where:
an effective area segmentation module 301, configured to perform effective area segmentation processing on an initial tongue image to be processed to obtain a first tongue image and a first tongue contour mask image;
the image diffusion module 302 is configured to generate N first generated tongue images corresponding to the first tongue contour mask images based on a preset image diffusion model, where N is greater than or equal to 1;
The color correction module 303 is configured to perform color correction on the first tongue images based on N first generated tongue images, and generate N second tongue images corresponding to the N first generated tongue images one by one;
the image recognition module 304 is configured to take N second tongue images or a set of N second tongue images and the first tongue image as target image data, and call a preset image recognition model to recognize the target image data, so as to obtain a recognition result of the initial tongue image.
Further, the color correction module 303 specifically includes:
the characteristic point matching submodule is used for sequentially selecting one image from the first generated tongue image as a target generated tongue image, and carrying out characteristic point matching on the first tongue image and the target generated tongue image to obtain a characteristic point pair;
the matrix generation sub-module is used for generating a color correction matrix between the first tongue image and the target generation type tongue image according to the color corresponding relation of the characteristic points;
and the correction calculation sub-module is used for carrying out color correction on the first tongue image based on the color correction matrix so as to obtain a second tongue image corresponding to the target generation type tongue image.
The image recognition module 304 of one embodiment of the present invention includes:
the text description prediction sub-module is used for inputting each piece of target image data into the text description prediction model for recognition so as to obtain text descriptions corresponding to each piece of target image data, wherein the text descriptions are used for describing tongue texture characteristics in the target image data;
and the tongue image classification prediction sub-module is used for inputting text descriptions corresponding to each target image data into the tongue image classification model for recognition so as to obtain a recognition result corresponding to the initial tongue image.
Further, the tongue image recognition device based on image augmentation further comprises a text description prediction model training module for training the text description prediction model in advance.
The text description prediction model training module specifically comprises:
the first acquisition submodule is used for carrying out effective area segmentation processing on the first initial tongue image sample to obtain a first tongue image sample;
a first obtaining sub-module for obtaining a text description describing tongue texture features of the first tongue image sample;
the text description training sub-module is used for inputting the first tongue image sample and the text description of the first tongue image sample as training data into an input layer of a preset multi-mode training model, and obtaining a text description prediction model through training.
Further, the tongue image recognition device based on image augmentation according to the embodiment of the invention further comprises a tongue image classification model training module for training the tongue image classification model in advance.
The tongue image classification model training module of one embodiment of the invention specifically comprises:
the second acquisition sub-module is used for carrying out effective area segmentation processing on a second initial tongue image sample to obtain a second tongue image sample and a first tongue outline mask image sample, and carrying out tongue image classification labeling on the second initial tongue image sample;
the image diffusion submodule is used for generating N first generation type tongue image samples corresponding to the second tongue outline mask image based on a preset image diffusion model;
the color correction sub-module is used for respectively carrying out color correction on the second tongue image samples based on N first generation type tongue image samples, and respectively generating N third tongue image samples which are in one-to-one correspondence with the N first generation type tongue image samples;
the text description acquisition sub-module is used for taking each third tongue image sample or a set of each third tongue image sample and the first tongue image sample as a target image data sample, and calling a preset text description prediction model to learn and identify the target image data sample so as to obtain text description corresponding to each target image data sample;
The first tongue image classification training sub-module is used for inputting a tongue image classification result of the second initial tongue image sample and text descriptions corresponding to each target image data sample as training data into an input layer of a preset tongue image classification prediction training model, and obtaining a tongue image classification model through training.
The tongue image classification model training module of another embodiment of the invention specifically comprises:
the third acquisition sub-module is used for acquiring text descriptions of a plurality of tongue texture features corresponding to different tongue image classification results;
the second tongue image classification training sub-module is used for inputting any target tongue image classification result and text description of a plurality of tongue texture features corresponding to the current target tongue image classification result as training data into an input layer of a preset tongue image classification prediction training model, and obtaining a tongue image classification model through training.
The image recognition module 304 of another embodiment of the present invention includes:
the image recognition sub-module is used for inputting each piece of target image data into the image recognition model for recognition so as to obtain tongue image classification results corresponding to each piece of target image data;
and the statistical calculation sub-module is used for carrying out statistical calculation on tongue image classification results corresponding to each target image data so as to finally obtain an initial tongue image recognition result.
Further, the tongue image recognition device based on image augmentation according to the embodiment of the invention further comprises a tongue segmentation model training module for training the tongue segmentation model in advance.
The tongue segmentation model training module specifically comprises:
a fourth acquisition sub-module for acquiring a third initial tongue image sample comprising a tongue active area;
the labeling sub-module is used for labeling the tongue effective areas in each third initial tongue image sample to obtain semantic segmentation training samples;
the semantic segmentation training sub-module is used for inputting the semantic segmentation training sample into an input layer of a preset semantic segmentation model, and obtaining a tongue segmentation model through training.
Further, the effective area segmentation module 301 of the embodiment of the present invention is further configured to determine whether the initial tongue image includes a tongue effective area, and if not, send an error prompt.
Further, the tongue image recognition device based on image augmentation according to the embodiment of the invention further comprises an image diffusion model training module for training the image diffusion model in advance.
The image diffusion model training module specifically comprises:
a fifth acquisition sub-module, configured to acquire a fourth initial tongue image sample that can truly reflect the texture characteristics of the tongue;
The effective area segmentation submodule is used for carrying out effective area segmentation processing on the fourth initial tongue image sample to obtain a fourth tongue image sample and a second tongue outline mask image sample;
and the image diffusion training sub-module is used for inputting the fourth tongue image sample and the second tongue outline mask image sample as training data into an input layer of a preset image diffusion training model, and obtaining the image diffusion model through training.
For the device embodiments, since they are substantially similar to the method embodiments, the description is relatively simple, and reference is made to the description of the method embodiments for relevant points.
The apparatus embodiments described above are merely illustrative, wherein the elements illustrated as separate elements may or may not be physically separate, and the elements shown as elements may or may not be physical elements, may be located in one place, or may be distributed over a plurality of network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of this embodiment. Those of ordinary skill in the art will understand and implement the present invention without undue burden.
After an initial tongue image is obtained, a first tongue image and a first tongue outline mask image which are subjected to effective area segmentation processing on the initial tongue image are obtained; generating N first generation type tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, respectively carrying out color correction on the first tongue images based on the N first generation type tongue images, respectively generating N second tongue images corresponding to the N first generation type tongue images one by one, and improving generalization capability and robustness in image recognition by acquiring the N second tongue images after color correction; and finally, taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image.
Example III
The embodiment of the invention provides a computer device, which comprises a memory, a processor and a computer program stored on the memory and capable of running on the processor, wherein the processor realizes the steps in the tongue image identification method embodiments based on image augmentation, such as the steps S1-S4 shown in figure 1 when executing the computer program. Alternatively, the processor, when executing the computer program, implements the functions of the modules/units in the tongue image recognition device or gateway system embodiments based on image augmentation, such as the effective area segmentation module 301, the image diffusion module 302, the color correction module 303, and the image recognition module 304 shown in fig. 3.
Furthermore, those skilled in the art will appreciate that while some embodiments herein include some features but not others included in other embodiments, combinations of features of different embodiments are meant to be within the scope of the invention and form different embodiments. For example, any of the claimed embodiments can be used in any combination.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention, and are not limiting; although the invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit and scope of the technical solutions of the embodiments of the present invention.

Claims (11)

1. A tongue image recognition method based on image augmentation, the method comprising:
carrying out effective region segmentation processing on an initial tongue image to be processed to obtain a first tongue image and a first tongue outline mask image;
generating N first generated tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, wherein N is greater than or equal to 1;
respectively carrying out color correction on the first tongue images based on N first generation type tongue images, and respectively generating N second tongue images corresponding to the N first generation type tongue images one by one;
and taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image.
2. The method according to claim 1, wherein the performing color correction on the first tongue images based on the N first generated tongue images respectively, and generating N second tongue images corresponding to the N first generated tongue images one to one respectively includes:
sequentially selecting one image from the first generated tongue image as a target generated tongue image, and performing feature point matching on the first tongue image and the target generated tongue image to obtain feature point pairs;
Generating a color correction matrix between the first tongue image and the target generated tongue image according to the color corresponding relation of the feature points;
and performing color correction on the first tongue image based on the color correction matrix to obtain a second tongue image corresponding to the target generated tongue image.
3. The method of claim 1, wherein the image recognition model is a cascading network model, a first network model of the cascading network model is a text description prediction model, and a second network model of the cascading network model is a tongue image classification prediction model;
the step of calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image comprises the following steps:
inputting each piece of target image data into the text description prediction model for recognition so as to obtain text descriptions corresponding to each piece of target image data, wherein the text descriptions are used for describing tongue texture features in the target image data;
and inputting text descriptions corresponding to the target image data into the tongue image classification prediction model for recognition so as to obtain a recognition result corresponding to the initial tongue image.
4. A method according to claim 3, wherein before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further comprises: pre-training the text description prediction model;
the pre-training the text description prediction model specifically comprises the following steps:
carrying out effective area segmentation processing on the first initial tongue image sample to obtain a first tongue image sample;
acquiring a text description describing tongue texture features of the first tongue image sample;
and inputting the first tongue image sample and the text description of the first tongue image sample as training data into an input layer of a preset multi-mode training model, and obtaining a text description prediction model through training.
5. The method of claim 4, wherein before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further comprises: pre-training the tongue image classification prediction model;
the pre-training the tongue image classification prediction model specifically comprises the following steps:
carrying out effective region segmentation processing on a second initial tongue image sample to obtain a second tongue image sample and a first tongue outline mask image sample, and carrying out tongue image classification labeling on the second initial tongue image sample;
Generating N first generation type tongue image samples corresponding to the second tongue outline mask image based on a preset image diffusion model;
respectively carrying out color correction on the second tongue image samples based on N first generation type tongue image samples, and respectively generating N third tongue image samples which are in one-to-one correspondence with the N first generation type tongue image samples;
taking N third tongue image samples or a set of N third tongue image samples and second tongue image samples as target image data samples, and calling a preset text description prediction model to identify the target image data samples so as to obtain text descriptions corresponding to the target image data samples;
and inputting a tongue image classification result of the second initial tongue image sample and text descriptions corresponding to each target image data sample as training data into an input layer of a preset tongue image classification prediction training model, and obtaining the tongue image classification prediction model through training.
6. A method according to claim 3, wherein before invoking a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image, the method further comprises: pre-training the tongue image classification prediction model;
The pre-training the tongue image classification prediction model specifically comprises the following steps:
acquiring text descriptions of a plurality of tongue texture features corresponding to different tongue image classification results;
and inputting a target tongue image classification result and text descriptions of a plurality of tongue texture features corresponding to the current target tongue image classification as training data into an input layer of a preset tongue image classification prediction training model, and obtaining the tongue image classification prediction model through training.
7. The method according to claim 1, wherein the calling a preset image recognition model to recognize the target image data to obtain a recognition result of the initial tongue image includes:
inputting each target image data into the image recognition model for recognition so as to obtain tongue image classification results corresponding to each target image data;
and carrying out statistical calculation on tongue image classification results corresponding to each target image data so as to finally obtain an initial tongue image recognition result.
8. The method according to any one of claims 1-7, wherein before performing the active area segmentation process on the initial tongue image to be processed to obtain the first tongue image and the first tongue contour mask image, the method further comprises: pre-training a tongue segmentation model;
The pre-trained tongue segmentation model comprises:
acquiring a third initial tongue image sample containing a tongue effective area;
labeling the tongue effective areas in each third initial tongue image sample to obtain a semantic segmentation training sample;
inputting the semantic segmentation training sample into an input layer of a preset semantic segmentation model, and obtaining a tongue segmentation model through training.
9. The method according to any one of claims 1-7, wherein before performing an active area segmentation process on an initial tongue image to be processed to obtain a first tongue image and a first tongue contour mask image, the method further comprises;
judging whether the initial tongue image contains a tongue effective area, and if not, sending out an error prompt.
10. A tongue image recognition device based on image augmentation, the device comprising:
the effective area segmentation module is used for carrying out effective area segmentation processing on the initial tongue image to be processed to obtain a first tongue image and a first tongue outline mask image;
the image diffusion module is used for generating N first generated tongue images corresponding to the first tongue outline mask images based on a preset image diffusion model, wherein N is greater than or equal to 1;
The color correction module is used for respectively carrying out color correction on the first tongue images based on N first generation type tongue images and respectively generating N second tongue images which are in one-to-one correspondence with the N first generation type tongue images;
the image recognition module is used for taking N second tongue images or a set of N second tongue images and the first tongue images as target image data, and calling a preset image recognition model to recognize the target image data so as to obtain a recognition result of the initial tongue image.
11. A computer device comprising a memory, a processor, and a computer program stored on the memory and executable on the processor; the computer program, when executed by the processor, implements the tongue image recognition method based on image augmentation as claimed in any one of claims 1-9.
CN202311051530.4A 2023-08-21 2023-08-21 Tongue image identification method and device based on image amplification and computer equipment Active CN117094966B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311051530.4A CN117094966B (en) 2023-08-21 2023-08-21 Tongue image identification method and device based on image amplification and computer equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311051530.4A CN117094966B (en) 2023-08-21 2023-08-21 Tongue image identification method and device based on image amplification and computer equipment

Publications (2)

Publication Number Publication Date
CN117094966A true CN117094966A (en) 2023-11-21
CN117094966B CN117094966B (en) 2024-04-05

Family

ID=88781442

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311051530.4A Active CN117094966B (en) 2023-08-21 2023-08-21 Tongue image identification method and device based on image amplification and computer equipment

Country Status (1)

Country Link
CN (1) CN117094966B (en)

Citations (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001186365A (en) * 1999-12-27 2001-07-06 Canon Inc Picture processing method, picture processor and recording medium
KR20040059312A (en) * 2002-12-28 2004-07-05 삼성전자주식회사 Method of extracting tongue area from tongue image and health care service method and apparatus using tongue image
WO2010002070A1 (en) * 2008-06-30 2010-01-07 Korea Institute Of Oriental Medicine Method for grouping 3d models to classify constitution
CN102509312A (en) * 2011-09-20 2012-06-20 哈尔滨工业大学 Color range space of human body digital tongue image color and extraction method thereof
KR20130083613A (en) * 2012-01-13 2013-07-23 상지대학교산학협력단 Method and apparatus for processing image of tongue using three band image
KR101373471B1 (en) * 2012-10-15 2014-03-13 세종대학교산학협력단 Apparatus and method for compensation of stereo image
JP2016103759A (en) * 2014-11-28 2016-06-02 株式会社リコー Image processing apparatus, image processing method, and program
US20170118407A1 (en) * 2015-10-22 2017-04-27 Samsung Electronics Co., Ltd. Method and device for generating images
KR101731243B1 (en) * 2015-12-15 2017-04-28 군산대학교 산학협력단 A video surveillance apparatus for identification and tracking multiple moving objects with similar colors and method thereof
KR20170099066A (en) * 2016-02-23 2017-08-31 정종율 Method and system for tongue diagnosis based on image of tongue
US20180095533A1 (en) * 2016-09-30 2018-04-05 Samsung Electronics Co., Ltd. Method for displaying an image and an electronic device thereof
CN109461128A (en) * 2018-10-24 2019-03-12 福州大学 A kind of consistent stereo-picture color calibration method of structure kept based on gradient
WO2019084919A1 (en) * 2017-11-03 2019-05-09 SZ DJI Technology Co., Ltd. Methods and system for infrared tracking
CN109978873A (en) * 2019-03-31 2019-07-05 山西慧虎健康科技有限公司 A kind of intelligent physical examination system and method based on Chinese medicine image big data
WO2019137131A1 (en) * 2018-01-10 2019-07-18 Oppo广东移动通信有限公司 Image processing method, apparatus, storage medium, and electronic device
WO2019218826A1 (en) * 2018-05-17 2019-11-21 腾讯科技(深圳)有限公司 Image processing method and device, computer apparatus, and storage medium
CN110598533A (en) * 2019-07-31 2019-12-20 平安科技(深圳)有限公司 Tongue picture matching method, electronic device, computer device, and storage medium
JP2020009162A (en) * 2018-07-09 2020-01-16 キヤノン株式会社 Image processing device, image processing method and program
KR20200092492A (en) * 2019-01-11 2020-08-04 연세대학교 산학협력단 Method and Apparatus for Image Adjustment Based on Semantics-Aware
CN111639647A (en) * 2020-05-22 2020-09-08 深圳市赛为智能股份有限公司 Indicating lamp state identification method and device, computer equipment and storage medium
WO2020188794A1 (en) * 2019-03-20 2020-09-24 株式会社日立国際電気 Video system, imaging device, and video processing device
CN112839216A (en) * 2021-01-13 2021-05-25 合肥埃科光电科技有限公司 Image color correction method and device
CN112884682A (en) * 2021-01-08 2021-06-01 福州大学 Stereo image color correction method and system based on matching and fusion
WO2021139258A1 (en) * 2020-06-19 2021-07-15 平安科技(深圳)有限公司 Image recognition based cell recognition and counting method and apparatus, and computer device
WO2021169325A1 (en) * 2020-02-25 2021-09-02 苏州科达科技股份有限公司 Gaze adjustment method and apparatus, and storage medium
CN113724228A (en) * 2021-08-31 2021-11-30 平安科技(深圳)有限公司 Tongue color and coating color identification method and device, computer equipment and storage medium
CN113781468A (en) * 2021-09-23 2021-12-10 河南科技大学 Tongue image segmentation method based on lightweight convolutional neural network
CN113780444A (en) * 2021-09-16 2021-12-10 平安科技(深圳)有限公司 Tongue fur image classification model training method based on progressive learning
CN113781488A (en) * 2021-08-02 2021-12-10 横琴鲸准智慧医疗科技有限公司 Tongue picture image segmentation method, apparatus and medium
CN113837986A (en) * 2020-12-15 2021-12-24 京东科技控股股份有限公司 Method, apparatus, electronic device, and medium for recognizing tongue picture
WO2022033150A1 (en) * 2020-08-11 2022-02-17 Oppo广东移动通信有限公司 Image recognition method, apparatus, electronic device, and storage medium
WO2022047662A1 (en) * 2020-09-02 2022-03-10 Intel Corporation Method and system of neural network object recognition for warpable jerseys with multiple attributes
WO2022055037A1 (en) * 2020-09-14 2022-03-17 안치영 Tongue diagnostic kit for color correction and saliva examination, and non-face-to-face tongue diagnostic system using same
CN114372926A (en) * 2021-12-21 2022-04-19 华东理工大学 Traditional Chinese medicine tongue tenderness identification method based on image restoration and convolutional neural network
CN114708493A (en) * 2022-02-26 2022-07-05 上海大学 Traditional Chinese medicine crack tongue diagnosis portable device and using method
CN114820603A (en) * 2022-06-27 2022-07-29 深圳中易健康科技有限公司 Intelligent health management method based on AI tongue diagnosis image processing and related device
WO2022247005A1 (en) * 2021-05-27 2022-12-01 平安科技(深圳)有限公司 Method and apparatus for identifying target object in image, electronic device and storage medium
CN115965607A (en) * 2022-12-31 2023-04-14 华东理工大学 Intelligent traditional Chinese medicine tongue diagnosis auxiliary analysis system
KR20230061757A (en) * 2021-10-29 2023-05-09 한국 한의학 연구원 Contact type measuring apparatus for papillas of tongue and coated tongue
CN116187470A (en) * 2023-01-18 2023-05-30 智慧眼科技股份有限公司 Tongue diagnosis image color correction model training method, color correction method and equipment
WO2023143129A1 (en) * 2022-01-30 2023-08-03 北京字跳网络技术有限公司 Image processing method and apparatus, electronic device, and storage medium

Patent Citations (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001186365A (en) * 1999-12-27 2001-07-06 Canon Inc Picture processing method, picture processor and recording medium
KR20040059312A (en) * 2002-12-28 2004-07-05 삼성전자주식회사 Method of extracting tongue area from tongue image and health care service method and apparatus using tongue image
WO2010002070A1 (en) * 2008-06-30 2010-01-07 Korea Institute Of Oriental Medicine Method for grouping 3d models to classify constitution
CN102509312A (en) * 2011-09-20 2012-06-20 哈尔滨工业大学 Color range space of human body digital tongue image color and extraction method thereof
KR20130083613A (en) * 2012-01-13 2013-07-23 상지대학교산학협력단 Method and apparatus for processing image of tongue using three band image
KR101373471B1 (en) * 2012-10-15 2014-03-13 세종대학교산학협력단 Apparatus and method for compensation of stereo image
JP2016103759A (en) * 2014-11-28 2016-06-02 株式会社リコー Image processing apparatus, image processing method, and program
US20170118407A1 (en) * 2015-10-22 2017-04-27 Samsung Electronics Co., Ltd. Method and device for generating images
KR101731243B1 (en) * 2015-12-15 2017-04-28 군산대학교 산학협력단 A video surveillance apparatus for identification and tracking multiple moving objects with similar colors and method thereof
KR20170099066A (en) * 2016-02-23 2017-08-31 정종율 Method and system for tongue diagnosis based on image of tongue
US20180095533A1 (en) * 2016-09-30 2018-04-05 Samsung Electronics Co., Ltd. Method for displaying an image and an electronic device thereof
WO2019084919A1 (en) * 2017-11-03 2019-05-09 SZ DJI Technology Co., Ltd. Methods and system for infrared tracking
WO2019137131A1 (en) * 2018-01-10 2019-07-18 Oppo广东移动通信有限公司 Image processing method, apparatus, storage medium, and electronic device
WO2019218826A1 (en) * 2018-05-17 2019-11-21 腾讯科技(深圳)有限公司 Image processing method and device, computer apparatus, and storage medium
JP2020009162A (en) * 2018-07-09 2020-01-16 キヤノン株式会社 Image processing device, image processing method and program
CN109461128A (en) * 2018-10-24 2019-03-12 福州大学 A kind of consistent stereo-picture color calibration method of structure kept based on gradient
KR20200092492A (en) * 2019-01-11 2020-08-04 연세대학교 산학협력단 Method and Apparatus for Image Adjustment Based on Semantics-Aware
WO2020188794A1 (en) * 2019-03-20 2020-09-24 株式会社日立国際電気 Video system, imaging device, and video processing device
CN109978873A (en) * 2019-03-31 2019-07-05 山西慧虎健康科技有限公司 A kind of intelligent physical examination system and method based on Chinese medicine image big data
CN110598533A (en) * 2019-07-31 2019-12-20 平安科技(深圳)有限公司 Tongue picture matching method, electronic device, computer device, and storage medium
WO2021017308A1 (en) * 2019-07-31 2021-02-04 平安科技(深圳)有限公司 Tongue image matching method, electronic apparatus, computer device, and storage medium
WO2021169325A1 (en) * 2020-02-25 2021-09-02 苏州科达科技股份有限公司 Gaze adjustment method and apparatus, and storage medium
CN111639647A (en) * 2020-05-22 2020-09-08 深圳市赛为智能股份有限公司 Indicating lamp state identification method and device, computer equipment and storage medium
WO2021139258A1 (en) * 2020-06-19 2021-07-15 平安科技(深圳)有限公司 Image recognition based cell recognition and counting method and apparatus, and computer device
WO2022033150A1 (en) * 2020-08-11 2022-02-17 Oppo广东移动通信有限公司 Image recognition method, apparatus, electronic device, and storage medium
WO2022047662A1 (en) * 2020-09-02 2022-03-10 Intel Corporation Method and system of neural network object recognition for warpable jerseys with multiple attributes
WO2022055037A1 (en) * 2020-09-14 2022-03-17 안치영 Tongue diagnostic kit for color correction and saliva examination, and non-face-to-face tongue diagnostic system using same
CN113837986A (en) * 2020-12-15 2021-12-24 京东科技控股股份有限公司 Method, apparatus, electronic device, and medium for recognizing tongue picture
CN112884682A (en) * 2021-01-08 2021-06-01 福州大学 Stereo image color correction method and system based on matching and fusion
CN112839216A (en) * 2021-01-13 2021-05-25 合肥埃科光电科技有限公司 Image color correction method and device
WO2022247005A1 (en) * 2021-05-27 2022-12-01 平安科技(深圳)有限公司 Method and apparatus for identifying target object in image, electronic device and storage medium
CN113781488A (en) * 2021-08-02 2021-12-10 横琴鲸准智慧医疗科技有限公司 Tongue picture image segmentation method, apparatus and medium
CN113724228A (en) * 2021-08-31 2021-11-30 平安科技(深圳)有限公司 Tongue color and coating color identification method and device, computer equipment and storage medium
CN113780444A (en) * 2021-09-16 2021-12-10 平安科技(深圳)有限公司 Tongue fur image classification model training method based on progressive learning
CN113781468A (en) * 2021-09-23 2021-12-10 河南科技大学 Tongue image segmentation method based on lightweight convolutional neural network
KR20230061757A (en) * 2021-10-29 2023-05-09 한국 한의학 연구원 Contact type measuring apparatus for papillas of tongue and coated tongue
CN114372926A (en) * 2021-12-21 2022-04-19 华东理工大学 Traditional Chinese medicine tongue tenderness identification method based on image restoration and convolutional neural network
WO2023143129A1 (en) * 2022-01-30 2023-08-03 北京字跳网络技术有限公司 Image processing method and apparatus, electronic device, and storage medium
CN114708493A (en) * 2022-02-26 2022-07-05 上海大学 Traditional Chinese medicine crack tongue diagnosis portable device and using method
CN114820603A (en) * 2022-06-27 2022-07-29 深圳中易健康科技有限公司 Intelligent health management method based on AI tongue diagnosis image processing and related device
CN115965607A (en) * 2022-12-31 2023-04-14 华东理工大学 Intelligent traditional Chinese medicine tongue diagnosis auxiliary analysis system
CN116187470A (en) * 2023-01-18 2023-05-30 智慧眼科技股份有限公司 Tongue diagnosis image color correction model training method, color correction method and equipment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZHUANG Q等: "Human-computer interaction based health diagnostics using ResNet34 for tongue image classification", 《COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE》, vol. 226, pages 1 - 16 *
李家炜: "基于深度学习的中医舌象特征分类方法研究", 《中国优秀硕士学位论文全文数据库 医药卫生科技辑》, no. 1, pages 056 - 25 *

Also Published As

Publication number Publication date
CN117094966B (en) 2024-04-05

Similar Documents

Publication Publication Date Title
CN110505459B (en) Image color correction method, device and storage medium suitable for endoscope
CN108960232A (en) Model training method, device, electronic equipment and computer readable storage medium
WO2021218119A1 (en) Image toning enhancement method and method for training image toning enhancement neural network
CN110502986A (en) Identify character positions method, apparatus, computer equipment and storage medium in image
CN108921161A (en) Model training method, device, electronic equipment and computer readable storage medium
CN108510560B (en) Image processing method, image processing device, storage medium and computer equipment
CN113743384B (en) Stomach picture identification method and device
CN111179252B (en) Cloud platform-based digestive tract disease focus auxiliary identification and positive feedback system
CN109360254A (en) Image processing method and device, electronic equipment, computer readable storage medium
CN111488912B (en) Laryngeal disease diagnosis system based on deep learning neural network
CN110956080A (en) Image processing method and device, electronic equipment and storage medium
CN112836653A (en) Face privacy method, device and apparatus and computer storage medium
CN110047059B (en) Image processing method and device, electronic equipment and readable storage medium
CN107920205A (en) Image processing method, device, storage medium and electronic equipment
CN112926508B (en) Training method and device of living body detection model
CN117094966B (en) Tongue image identification method and device based on image amplification and computer equipment
CN113052768B (en) Method, terminal and computer readable storage medium for processing image
CN110110750B (en) Original picture classification method and device
CN112381073A (en) IQ (in-phase/quadrature) adjustment method and adjustment module based on AI (Artificial Intelligence) face detection
CN110097080B (en) Construction method and device of classification label
CN114972065A (en) Training method and system of color difference correction model, electronic equipment and mobile equipment
CN112529002B (en) Tongue picture classification method and device, computer equipment and storage medium
CN107948619B (en) Image processing method, device, computer readable storage medium and mobile terminal
CN108881740A (en) Image method and device, electronic equipment, computer readable storage medium
CN111767829B (en) Living body detection method, device, system and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant