WO2022178934A1 - Health testing method and apparatus, and device and storage medium - Google Patents

Health testing method and apparatus, and device and storage medium Download PDF

Info

Publication number
WO2022178934A1
WO2022178934A1 PCT/CN2021/082864 CN2021082864W WO2022178934A1 WO 2022178934 A1 WO2022178934 A1 WO 2022178934A1 CN 2021082864 W CN2021082864 W CN 2021082864W WO 2022178934 A1 WO2022178934 A1 WO 2022178934A1
Authority
WO
WIPO (PCT)
Prior art keywords
health
voice
data
training
tongue
Prior art date
Application number
PCT/CN2021/082864
Other languages
French (fr)
Chinese (zh)
Inventor
顾艳梅
马骏
王少军
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2022178934A1 publication Critical patent/WO2022178934A1/en

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/30ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for calculating health indices; for individual health risk assessment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/254Fusion techniques of classification results, e.g. of results related to same input data
    • G06F18/256Fusion techniques of classification results, e.g. of results related to same input data of results relating to different input data, e.g. multimodal recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/66Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition

Definitions

  • the present application relates to the field of artificial intelligence, and in particular, to a health detection method, apparatus, electronic device, and computer-readable storage medium.
  • a health detection method provided by this application includes:
  • the voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  • the present application also provides a health detection device, the device comprising:
  • the acquisition module is used to acquire the user's voice data and tongue image
  • a detection module configured to perform voice health detection on the voice data by using the trained voice health recognition model to obtain a voice health detection result
  • the detection module is used to perform image health detection on the tongue image by using the trained tongue health recognition model to obtain a tongue health detection result;
  • a push module is configured to fuse the voice health detection result and the tongue health detection result to obtain a health detection result, and push the health detection result to the user.
  • the present application also provides an electronic device, the electronic device comprising:
  • the memory stores a computer program executable by the at least one processor, the computer program being executed by the at least one processor to implement the following steps:
  • the voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  • the present application also provides a computer-readable storage medium, where at least one computer program is stored in the computer-readable storage medium, and the at least one computer program is executed by a processor in an electronic device to implement the following steps:
  • the voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  • FIG. 1 is a schematic flowchart of a health detection method provided by an embodiment of the present application.
  • FIG. 2 is a detailed flowchart of one step of the health detection method provided in FIG. 1 in the first embodiment of the present application;
  • FIG. 3 is a schematic block diagram of a health detection device provided by an embodiment of the present application.
  • FIG. 4 is a schematic diagram of the internal structure of an electronic device implementing a health detection method provided by an embodiment of the present application
  • the embodiments of the present application provide a health detection method.
  • the execution body of the health detection method includes, but is not limited to, at least one of electronic devices that can be configured to execute the method provided by the embodiments of the present application, such as a server and a terminal.
  • the health detection method can be executed by software or hardware installed in a terminal device or a server device, and the software can be a blockchain platform.
  • the server includes but is not limited to: a single server, a server cluster, a cloud server or a cloud server cluster, and the like.
  • the health detection method includes:
  • the embodiment of the present application obtains the user's voice data and tongue image to ensure the premise of subsequent user health detection. This can help users better understand their own physique.
  • the voice data refers to the voice data issued by the user
  • the tongue image refers to a picture of the user's tongue.
  • the voice data may be acquired through a sound collection device, and the sound collection device includes a mobile phone microphone.
  • the tongue phase image can be acquired by an image acquisition device, and the image acquisition device includes a mobile phone camera.
  • the voice health recognition model includes a voice classification module and a voice analysis module, wherein the voice classification module is used to extract user voices from the voice data, so as to segment the background sound of the voice data , extracting the user's voice, and the voice analysis module is configured to perform a voice health analysis on the user's voice output by the voice classification module, so as to detect the voice health detection result of the user.
  • the voice health recognition model needs to be trained to ensure the voice health of the voice health recognition model. detection accuracy.
  • the training of the voice health recognition model includes:
  • the human voice data and health detection in S20 can be implemented by manual labeling, so as to ensure the accuracy of the generated standard human voice data and standard human voice health data, so that the Better supervise the learning ability of subsequent models.
  • the use of the voice classification module of the voice health recognition model to perform human voice segmentation on the training voice data to obtain training voice data includes: using the voice classification module in the voice classification module.
  • the voice frequency conversion algorithm converts the training voice data into corresponding voice frequencies, calculates the dimensional parameters of the voice frequencies, and filters out the human voice data in the training voice data according to the dimensional parameters to obtain the training voice.
  • the dimension parameters include: intonation, speech rate, and the like. For example, convert a user's voice into a voice frequency in the range of 70-100 Hz, calculate the user's intonation, speed and other dimension parameters according to the voice frequency, and filter out the people in the training voice data according to the dimension parameters.
  • the sound frequency conversion algorithm includes:
  • B(f) represents the speech frequency and f represents the expected frequency of the training speech data.
  • the following method is used to calculate the dimension parameter of the voice frequency:
  • d(n) represents the dimension parameter of the speech frequency
  • i represents the frame rate of the speech frequency
  • n represents the amplitude of the speech frequency
  • B(f) represents the speech frequency
  • k represents the linear combination of the current standard speech frame and the preceding and following standard speech frames , usually takes a value of 2, which represents the linear combination of the current speech frame and the two preceding and following speech frames.
  • the step S22 includes: using the speech analysis module to perform feature extraction on the training vocal data to obtain characteristic speech data, and performing health analysis on the characteristic speech data and outputting, Obtain the training vocal health data.
  • the characteristic voice data refers to the characteristic voiceprint in the training voice data, which is used to represent the voice information of the training voice data, and the health analysis may Frequency and other dimensional information to establish convolution kernel implementation.
  • the S23 includes: calculating a first loss value of the voice health recognition model according to the standard human voice data and the training human voice data; The health data and the training vocal health data are used to calculate the second loss value of the voice health recognition model; the loss value of the voice health recognition model is calculated according to the first loss value and the second loss value.
  • the following method is used to calculate the first loss value of the voice health recognition model:
  • L(s) represents the first loss value
  • k represents the number of training speech data
  • yi represents the ith training vocal data
  • y′ i represents the ith standard human voice data.
  • the second loss value of the voice health recognition model is calculated by the following method:
  • L1 represents the second loss value
  • ⁇ g represents the standard vocal health data
  • ⁇ p is the training vocal health data
  • the preset condition includes that the loss value is less than a loss threshold. That is, when the loss value is less than the loss threshold, it means that the loss value satisfies the preset condition, and when the loss value is greater than or equal to the loss threshold, it means that the loss value does not meet the predetermined condition. when the preset conditions are stated.
  • the loss threshold may be set to 0.1, or may be set according to actual scenarios.
  • the parameter adjustment of the voice health recognition model may be implemented by a currently known stochastic gradient descent algorithm, which will not be described further herein.
  • the tongue health recognition model includes an image classification module and an image analysis module, wherein the image classification module is used to perform background segmentation on the tongue image to output the tongue region, the image The analysis module is configured to perform a tongue health analysis on the tongue region output by the image classification module, and output the tongue health state.
  • the tongue health recognition model needs to be trained to ensure the tongue health recognition.
  • the model's tongue health detection accuracy is the case of the present application.
  • the training of the tongue health recognition model includes: acquiring a training tongue image, and using an image classification module in the tongue health recognition model to perform feature extraction on the training tongue image to obtain features Tongue image, use the image analysis module in the tongue health recognition model to detect the health status of the characteristic tongue image, obtain the predicted health status, and calculate the predicted health status and the standard health corresponding to the training tongue image
  • the training loss of the state, according to the training loss, the parameters of the tongue health recognition model are adjusted until the training loss is less than the preset training loss, and the trained tongue health recognition model is obtained.
  • the preset training loss is 0.1.
  • performing feature extraction on the training lingual image by using the image classification module in the lingual health recognition model to obtain a characteristic lingual image includes: using the image classification The convolution layer in the module performs the convolution operation on the training tongue image to obtain the initial characteristic tongue image, and uses the pooling layer in the image classification module to perform a dimensionality reduction operation on the initial characteristic tongue image to obtain The dimension-reduced characteristic lingual image is outputted by using the activation function in the image classification module to obtain the characteristic lingual image.
  • the activation function in the image classification module includes a relu activation function.
  • the detecting the health state of the characteristic tongue image by using the image analysis module in the tongue health recognition model to obtain the predicted health state includes: using the image analysis module The sampling layer in the sample up-samples the characteristic lingual image to obtain the sampled lingual image, and uses the fully connected layer in the image analysis module to perform health detection on the sampled lingual image and output it to obtain the prediction. health status.
  • the training loss may be calculated by a currently known sigmoid function.
  • the method before performing image health detection on the lingual image by using the trained lingual health recognition model, the method further includes: performing a preprocessing operation on the lingual image to improve the lingual image.
  • the quality of the images ensures the accuracy of the analysis of lingual images.
  • the preprocessing operation includes: performing a grayscale conversion operation on the tongue image through each proportional method to obtain a grayscale tongue image; using Gaussian filtering to reduce noise on the grayscale tongue image; and using contrast
  • the enhancement is to perform contrast enhancement on the noise-reduced grayscale lingual image; and perform a thresholding operation on the contrast-enhanced grayscale lingual image according to the OTSU algorithm.
  • f(x, a) represents the health detection result
  • k represents the number of the fused health detection results
  • x represents the same vector of the voice health detection result and the tongue health detection result
  • x represents the same vector of the voice health detection result and the tongue health detection result
  • represents the preset weight parameter (a ⁇ (0,1)).
  • the division of the weight parameters can be implemented according to actual business scenarios, for example, the weight of the tongue health detection result is 60%, and the weight of the voice health detection result is 40%.
  • the health detection result can be pushed to the user through a mobile terminal, so that the user can intuitively understand his physical state in real time, wherein the mobile terminal can be a mobile terminal.
  • the health detection results can also be stored in a blockchain node.
  • the embodiment of the present application first obtains the user's voice data and tongue image to ensure the premise of subsequent user health detection; secondly, the embodiment of the present application uses the trained voice health recognition model to perform voice health detection on the voice data to obtain a voice Health detection results, and use the trained tongue health recognition model to perform image health detection on the tongue images to obtain tongue health detection results, which can realize the health detection of online users' voice data and tongue data, so that users can The health detection is more comprehensive; further, the embodiment of the present application fuses the voice health detection result and the tongue health detection result to obtain a health detection result, and pushes the health detection result to the user to Help users understand their physical state intuitively in real time. Therefore, the present application can improve the convenience of user health detection.
  • FIG. 3 it is a functional block diagram of the health detection device of the present application.
  • the health detection apparatus 100 described in this application may be installed in an electronic device. According to the implemented functions, the health detection apparatus may include an acquisition module 101 , a detection module 102 and a push module 103 .
  • the modules described in the present invention can also be called units, which refer to a series of computer program segments that can be executed by the electronic device processor and can perform fixed functions, and are stored in the memory of the electronic device.
  • each module/unit is as follows:
  • the acquisition module 101 is used to acquire the user's voice data and tongue image
  • the detection module 102 is configured to perform voice health detection on the voice data by using the trained voice health recognition model to obtain a voice health detection result;
  • the detection module 102 is configured to perform image health detection on the tongue image by using the trained tongue health recognition model to obtain a tongue health detection result;
  • the pushing module 103 is configured to fuse the voice health detection result and the tongue health detection result to obtain a health detection result, and push the health detection result to the user.
  • the modules in the health detection device 100 in the embodiments of the present application use the same technical means as the health detection methods described in the above-mentioned FIG. 1 and FIG. 2 , and can generate the same technology The effect will not be repeated here.
  • FIG. 4 it is a schematic structural diagram of an electronic device implementing the health detection method of the present application.
  • the electronic device 1 may include a processor 10, a memory 11 and a bus, and may also include a computer program stored in the memory 11 and executable on the processor 10, such as a health detection program 12.
  • the memory 11 includes at least one type of readable storage medium, and the readable storage medium includes flash memory, mobile hard disk, multimedia card, card-type memory (for example: SD or DX memory, etc.), magnetic memory, magnetic disk, CD etc.
  • the memory 11 may be an internal storage unit of the electronic device 1 in some embodiments, such as a mobile hard disk of the electronic device 1 .
  • the memory 11 may also be an external storage device of the electronic device 1, such as a pluggable mobile hard disk, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital) equipped on the electronic device 1. , SD) card, flash memory card (Flash Card), etc.
  • the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device.
  • the memory 11 can not only be used to store application software installed in the electronic device 1 and various types of data, such as codes of the health detection program 12, etc., but also can be used to temporarily store data that has been output or will be output.
  • the processor 10 may be composed of integrated circuits, for example, may be composed of a single packaged integrated circuit, or may be composed of multiple integrated circuits packaged with the same function or different functions, including one or more integrated circuits.
  • Central Processing Unit CPU
  • microprocessor digital processing chip
  • graphics processor and combination of various control chips, etc.
  • the processor 10 is the control core (Control Unit) of the electronic device, and uses various interfaces and lines to connect the various components of the entire electronic device, by running or executing the program or module (for example, executing the program) stored in the memory 11. health detection program, etc.), and call the data stored in the memory 11 to perform various functions of the electronic device 1 and process data.
  • the bus may be a peripheral component interconnect (PCI for short) bus or an extended industry standard architecture (Extended industry standard architecture, EISA for short) bus or the like.
  • PCI peripheral component interconnect
  • EISA Extended industry standard architecture
  • the bus can be divided into address bus, data bus, control bus and so on.
  • the bus is configured to implement connection communication between the memory 11 and at least one processor 10 and the like.
  • FIG. 4 only shows an electronic device with components. Those skilled in the art can understand that the structure shown in FIG. 4 does not constitute a limitation on the electronic device 1, and may include fewer or more components than those shown in the drawings. components, or a combination of certain components, or a different arrangement of components.
  • the electronic device 1 may also include a power supply (such as a battery) for powering the various components, preferably, the power supply may be logically connected to the at least one processor 10 through a power management device, so that the power management
  • the device implements functions such as charge management, discharge management, and power consumption management.
  • the power source may also include one or more DC or AC power sources, recharging devices, power failure detection circuits, power converters or inverters, power status indicators, and any other components.
  • the electronic device 1 may further include various sensors, Bluetooth modules, Wi-Fi modules, etc., which will not be repeated here.
  • the electronic device 1 may also include a network interface, optionally, the network interface may include a wired interface and/or a wireless interface (such as a WI-FI interface, a Bluetooth interface, etc.), which is usually used in the electronic device 1 Establish a communication connection with other electronic devices.
  • a network interface optionally, the network interface may include a wired interface and/or a wireless interface (such as a WI-FI interface, a Bluetooth interface, etc.), which is usually used in the electronic device 1 Establish a communication connection with other electronic devices.
  • the electronic device 1 may further include a user interface, and the user interface may be a display (Display), an input unit (eg, a keyboard (Keyboard)), optionally, the user interface may also be a standard wired interface or a wireless interface.
  • the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light-emitting diode) touch device, and the like.
  • the display may also be appropriately called a display screen or a display unit, which is used for displaying information processed in the electronic device 1 and for displaying a visualized user interface.
  • the health detection program 12 stored in the memory 11 in the electronic device 1 is a combination of multiple programs. When running in the processor 10, it can realize:
  • the voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  • the modules/units integrated in the electronic device 1 may be stored in a non-volatile computer-readable storage medium.
  • the computer-readable storage medium may be volatile or non-volatile.
  • the computer-readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a removable hard disk, a magnetic disk, an optical disc, a computer memory, a read-only memory (ROM, Read-Only). Memory).
  • the present application also provides a computer-readable storage medium, where the readable storage medium stores a computer program, and when executed by a processor of an electronic device, the computer program can realize:
  • the voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  • modules described as separate components may or may not be physically separated, and components shown as modules may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
  • each functional module in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware, or can be implemented in the form of hardware plus software function modules.
  • the blockchain referred to in this application is a new application mode of computer technologies such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm.
  • Blockchain essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information to verify its Validity of information (anti-counterfeiting) and generation of the next block.
  • the blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Physics & Mathematics (AREA)
  • Public Health (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Epidemiology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Pathology (AREA)
  • Databases & Information Systems (AREA)
  • Biomedical Technology (AREA)
  • Primary Health Care (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)

Abstract

A health testing method, relating to the field of artificial intelligence. The method comprises: acquiring voice data and a tongue appearance image of a user (S1); performing voice health testing on the voice data using a completely trained voice health recognition model, so as to obtain a voice health testing result (S2); performing image health testing on the tongue appearance image using a completely trained tongue appearance health recognition model, so as to obtain a tongue appearance health testing result (S3); and combining the voice health testing result with the tongue appearance health testing result to obtain a health testing result, and pushing the health testing result to the user (S4). By means of the method, health testing accuracy can be improved.

Description

健康检测方法、装置、设备及存储介质Health detection method, device, equipment and storage medium
本申请要求于2021年2月26日提交中国专利局、申请号为CN202110214173.3、名称为“健康检测方法、装置、设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number CN202110214173.3 and titled "Health Detection Method, Apparatus, Equipment and Storage Medium" filed with the China Patent Office on February 26, 2021, the entire contents of which are incorporated by reference in this application.
技术领域technical field
本申请涉及人工智能领域,尤其涉及一种健康检测方法、装置、电子设备及计算机可读存储介质。The present application relates to the field of artificial intelligence, and in particular, to a health detection method, apparatus, electronic device, and computer-readable storage medium.
背景技术Background technique
随着当前人民生活水平的不断提高,人们越来越关注自己的身体健康状况,同时人们期望现在的医疗技术能够更好的呵护到普通百姓,但是我国目前医疗系统方面还处在“看病难,看病贵”的阶段。发明人意识到,目前,关于用户健康检测主要是通过基于人脸识别身份,拉取对应的用户医疗数据,从而实现用户健康的检测,但是往往现实场景中,医疗数据有局限无法很全面了解到用户的全面医疗数据,从而会影响用户健康检测的准确性。With the continuous improvement of people's living standards, people are paying more and more attention to their physical health. At the same time, people expect that the current medical technology can better care for ordinary people. However, my country's current medical system is still in the "difficulty in seeing a doctor. Medical treatment is expensive” stage. The inventor realized that at present, the user health detection is mainly based on the face recognition identity, and the corresponding user medical data is pulled, so as to realize the user health detection, but often in real scenarios, the medical data is limited and cannot be fully understood. The comprehensive medical data of the user will affect the accuracy of the user's health detection.
发明内容SUMMARY OF THE INVENTION
为实现上述目的,本申请提供的一种健康检测方法,包括:In order to achieve the above purpose, a health detection method provided by this application includes:
获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
本申请还提供一种健康检测装置,所述装置包括:The present application also provides a health detection device, the device comprising:
获取模块,用于获取用户的语音数据和舌相图像;The acquisition module is used to acquire the user's voice data and tongue image;
检测模块,用于利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;a detection module, configured to perform voice health detection on the voice data by using the trained voice health recognition model to obtain a voice health detection result;
所述检测模块,用于利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;The detection module is used to perform image health detection on the tongue image by using the trained tongue health recognition model to obtain a tongue health detection result;
推送模块,用于对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。A push module is configured to fuse the voice health detection result and the tongue health detection result to obtain a health detection result, and push the health detection result to the user.
本申请还提供一种电子设备,所述电子设备包括:The present application also provides an electronic device, the electronic device comprising:
至少一个处理器;以及,at least one processor; and,
与所述至少一个处理器通信连接的存储器;其中,a memory communicatively coupled to the at least one processor; wherein,
所述存储器存储有可被所述至少一个处理器执行的计算机程序,所述计算机程序被所述至少一个处理器执行,以实现如下步骤:The memory stores a computer program executable by the at least one processor, the computer program being executed by the at least one processor to implement the following steps:
获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
本申请还提供一种计算机可读存储介质,所述计算机可读存储介质中存储有至少一个计算机程序,所述至少一个计算机程序被电子设备中的处理器执行以实现如下步骤:The present application also provides a computer-readable storage medium, where at least one computer program is stored in the computer-readable storage medium, and the at least one computer program is executed by a processor in an electronic device to implement the following steps:
获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
本申请的一个或多个实施例的细节在下面的附图和描述中提出,本申请的其他特征和优点将从说明书、附图以及权利要求变得明显。The details of one or more embodiments of the application are set forth in the accompanying drawings and the description below, and other features and advantages of the application will become apparent from the description, drawings, and claims.
附图说明Description of drawings
图1为本申请一实施例提供的健康检测方法的流程示意图;1 is a schematic flowchart of a health detection method provided by an embodiment of the present application;
图2为本申请第一实施例中图1提供的健康检测方法其中一个步骤的详细流程示意图;FIG. 2 is a detailed flowchart of one step of the health detection method provided in FIG. 1 in the first embodiment of the present application;
图3为本申请一实施例提供的健康检测装置的模块示意图;FIG. 3 is a schematic block diagram of a health detection device provided by an embodiment of the present application;
图4为本申请一实施例提供的实现健康检测方法的电子设备的内部结构示意图;4 is a schematic diagram of the internal structure of an electronic device implementing a health detection method provided by an embodiment of the present application;
本申请目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The realization, functional characteristics and advantages of the purpose of the present application will be further described with reference to the accompanying drawings in conjunction with the embodiments.
具体实施方式Detailed ways
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。It should be understood that the specific embodiments described herein are only used to explain the present application, but not to limit the present application.
本申请实施例提供一种健康检测方法。所述健康检测方法的执行主体包括但不限于服务端、终端等能够被配置为执行本申请实施例提供的该方法的电子设备中的至少一种。换言之,所述健康检测方法可以由安装在终端设备或服务端设备的软件或硬件来执行,所述软件可以是区块链平台。所述服务端包括但不限于:单台服务器、服务器集群、云端服务器或云端服务器集群等。The embodiments of the present application provide a health detection method. The execution body of the health detection method includes, but is not limited to, at least one of electronic devices that can be configured to execute the method provided by the embodiments of the present application, such as a server and a terminal. In other words, the health detection method can be executed by software or hardware installed in a terminal device or a server device, and the software can be a blockchain platform. The server includes but is not limited to: a single server, a server cluster, a cloud server or a cloud server cluster, and the like.
参照图1所示,为本申请一实施例提供的健康检测方法的流程示意图。在本申请实施例中,所述健康检测方法包括:Referring to FIG. 1 , a schematic flowchart of a health detection method according to an embodiment of the present application is shown. In the embodiment of the present application, the health detection method includes:
S1、获取用户的语音数据和舌相图像。S1. Acquire the voice data and the tongue image of the user.
应该了解,随着人们生活水平不断提高,人们越来越多的会关注自己身体健康状况,因此,本申请实施例通过获取用户的语音数据和舌相图像,以保障后续用户健康检测的前提,从而可以帮助用户更好的了解自身体质。其中,所述语音数据指的是用户发出的声音数据,所述舌相图像指的是用户舌相图片。It should be understood that with the continuous improvement of people's living standards, more and more people will pay attention to their physical health status. Therefore, the embodiment of the present application obtains the user's voice data and tongue image to ensure the premise of subsequent user health detection. This can help users better understand their own physique. Wherein, the voice data refers to the voice data issued by the user, and the tongue image refers to a picture of the user's tongue.
一个可选实施例中,所述语音数据可以通过声音采集设备获取,所述声音采集设备包括手机麦克风。In an optional embodiment, the voice data may be acquired through a sound collection device, and the sound collection device includes a mobile phone microphone.
一个可选实施例后,所述舌相图像可以通过图像采集设备获取,所述图像采集设备包括手机摄像头。After an optional embodiment, the tongue phase image can be acquired by an image acquisition device, and the image acquisition device includes a mobile phone camera.
S2、利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果。S2. Use the trained voice health recognition model to perform voice health detection on the voice data to obtain a voice health detection result.
本申请实施例中,所述语音健康识别模型包括语音分类模块和语音分析模块,其中,所述语音分类模块用于对所述语音数据进行用户声音提取,以分割出所述语音数据的背景声音,提取出用户声音,所述语音分析模块用于对所述语音分类模块输出的用户声音进行声音健康分析,以检测出所述用户的语音健康检测结果。In the embodiment of the present application, the voice health recognition model includes a voice classification module and a voice analysis module, wherein the voice classification module is used to extract user voices from the voice data, so as to segment the background sound of the voice data , extracting the user's voice, and the voice analysis module is configured to perform a voice health analysis on the user's voice output by the voice classification module, so as to detect the voice health detection result of the user.
进一步地,本申请实施例所述利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测之前,需要对所述语音健康识别模型进行训练,以保障所述语音健康识别模型的语音健康检测准确率。Further, before using the trained voice health recognition model to perform voice health detection on the voice data according to the embodiment of the present application, the voice health recognition model needs to be trained to ensure the voice health of the voice health recognition model. detection accuracy.
详细地,参阅图2所示,所述对所述语音健康识别模型进行训练,包括:In detail, referring to Fig. 2, the training of the voice health recognition model includes:
S20、获取训练语音数据,标记所述训练语音数据中的人声数据,得到标准人声数据,并对所述标准人声数据进行健康检测,得到标准人声健康数据;S20, acquiring training voice data, marking the human voice data in the training voice data to obtain standard human voice data, and performing health detection on the standard human voice data to obtain standard human voice health data;
S21、利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据;S21, using the voice classification module of the voice health recognition model to perform vocal segmentation on the training voice data to obtain training voice data;
S22、利用所述语音健康识别模型中的语音分析模块对所述训练人声数据进行健康检测,得到训练人声健康数据;S22, using the voice analysis module in the voice health recognition model to perform health detection on the training vocal data to obtain training vocal health data;
S23、根据所述标准人声数据、标准人声健康数据、训练人声数据以及所述训练人声健康数据,计算所述语音健康识别模型的损失值;S23, according to the standard vocal data, standard vocal health data, training vocal data and the training vocal health data, calculate the loss value of the voice health recognition model;
若所述损失值不满足预设条件,则执行S24、调整所述语音健康识别模型的参数,并返回所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据的步骤;If the loss value does not meet the preset condition, perform S24, adjust the parameters of the voice health recognition model, and return to the voice classification module using the voice health recognition model to perform vocal segmentation on the training voice data , the steps of obtaining training vocal data;
若所述损失值满足预设条件,则执行S25、得到训练完成的语音健康识别模型。If the loss value satisfies the preset condition, perform S25 to obtain a trained voice health recognition model.
在本申请的一个可选实施例中,所述S20中的人声数据及健康检测可以通过人工标注的方法实现,以确保生成的标准人声数据和标准人声健康数据的准确性,从而可以更好的监督后续模型的学习能力。In an optional embodiment of the present application, the human voice data and health detection in S20 can be implemented by manual labeling, so as to ensure the accuracy of the generated standard human voice data and standard human voice health data, so that the Better supervise the learning ability of subsequent models.
在本申请的一个可选实施例中,所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据,包括:利用所述语音分类模块中的声音频率转换算法将所述训练语音数据转换成对应的语音频率,计算所述语音频率的维度参数,根据所述维度参数,筛选出所述训练语音数据中的人声数据,得到训练人声数据。所述维度参数包括:语调、语速等。例如,将某用户的语音转换语音频率为70-100HZ范围内,根据其语音频率计算出该用户的语调、语速等维度参数,根据所述维度参数,筛选出所述训练语音数据中的人声数据In an optional embodiment of the present application, the use of the voice classification module of the voice health recognition model to perform human voice segmentation on the training voice data to obtain training voice data includes: using the voice classification module in the voice classification module. The voice frequency conversion algorithm converts the training voice data into corresponding voice frequencies, calculates the dimensional parameters of the voice frequencies, and filters out the human voice data in the training voice data according to the dimensional parameters to obtain the training voice. data. The dimension parameters include: intonation, speech rate, and the like. For example, convert a user's voice into a voice frequency in the range of 70-100 Hz, calculate the user's intonation, speed and other dimension parameters according to the voice frequency, and filter out the people in the training voice data according to the dimension parameters. sound data
一个可选实施例中,所述声音频率转换算法包括:In an optional embodiment, the sound frequency conversion algorithm includes:
Figure PCTCN2021082864-appb-000001
Figure PCTCN2021082864-appb-000001
其中,B(f)表示语音频率,f表示训练语音数据的预期频率。where B(f) represents the speech frequency and f represents the expected frequency of the training speech data.
一个可选实施例中,利用下述方法计算所述语音频率的维度参数:In an optional embodiment, the following method is used to calculate the dimension parameter of the voice frequency:
Figure PCTCN2021082864-appb-000002
Figure PCTCN2021082864-appb-000002
其中,d(n)表示语音频率的维度参数,i表示语音频率的帧率,n表示语音频率的振幅,B(f)表示语音频率,k表示当前标准语音帧与前后标准语音帧的线性组合,通常取值为2,表示当前语音帧与前后2个语音帧的线性组合。Among them, d(n) represents the dimension parameter of the speech frequency, i represents the frame rate of the speech frequency, n represents the amplitude of the speech frequency, B(f) represents the speech frequency, and k represents the linear combination of the current standard speech frame and the preceding and following standard speech frames , usually takes a value of 2, which represents the linear combination of the current speech frame and the two preceding and following speech frames.
在本申请的一个可选实施例中,所述S22包括:利用所述语音分析模块对所述训练人声数据进行特征提取,得到特征语音数据,对所述特征语音数据进行健康分析后输出,得到所述训练人声健康数据。其中,所述特征语音数据是指所述训练人声数据中的特征声纹,用于表征训练人声数据的语音信息,所述健康分析可以根据所述特征语音数据的语速、语调以及基频等维度信息建立卷积核实现。In an optional embodiment of the present application, the step S22 includes: using the speech analysis module to perform feature extraction on the training vocal data to obtain characteristic speech data, and performing health analysis on the characteristic speech data and outputting, Obtain the training vocal health data. The characteristic voice data refers to the characteristic voiceprint in the training voice data, which is used to represent the voice information of the training voice data, and the health analysis may Frequency and other dimensional information to establish convolution kernel implementation.
在本申请的一个可选实施例中,所述S23包括:根据所述标准人声数据和所述训练人声数据,计算所述语音健康识别模型的第一损失值;根据所述标准人声健康数据和所述训练人声健康数据,计算所述语音健康识别模型的第二损失值;根据所述第一损失值和所述第二损失值,计算所述语音健康识别模型的损失值。In an optional embodiment of the present application, the S23 includes: calculating a first loss value of the voice health recognition model according to the standard human voice data and the training human voice data; The health data and the training vocal health data are used to calculate the second loss value of the voice health recognition model; the loss value of the voice health recognition model is calculated according to the first loss value and the second loss value.
一个可选实施例中,利用下述方法计算所述语音健康识别模型的第一损失值:In an optional embodiment, the following method is used to calculate the first loss value of the voice health recognition model:
Figure PCTCN2021082864-appb-000003
Figure PCTCN2021082864-appb-000003
其中,L(s)表示第一损失值,k表示训练语音数据的数量,y i表示第i个训练人声数据,y′ i表示第i个标准人声数据。 Among them, L(s) represents the first loss value, k represents the number of training speech data, yi represents the ith training vocal data, and y′ i represents the ith standard human voice data.
一个可选实施例中,利用下述方法计算所述语音健康识别模型的第二损失值:In an optional embodiment, the second loss value of the voice health recognition model is calculated by the following method:
L1=|α pg| L1=|α pg |
其中L1表示第二损失值,α g表示标准人声健康数据,α p训练人声健康数据。 where L1 represents the second loss value, α g represents the standard vocal health data, and α p is the training vocal health data.
一个可选实施例中,所述根据所述第一训练损失和所述第二训练损失,计算所述语音健康识别模型的损失值,包括:将所述第一损失值和所述第二损失值进行相加,得到所述语音健康识别模型的损失值,即L=L(s)+LC。In an optional embodiment, the calculating the loss value of the speech health recognition model according to the first training loss and the second training loss includes: combining the first loss value and the second loss The values are added to obtain the loss value of the speech health recognition model, that is, L=L(s)+LC.
在本申请的一个可选实施例中,所述预设条件包括所述损失值小于损失阈值。即当所述损失值小于所述损失阈值时,则表示所述损失值满足所述预设条件时,当所述损失值大于或者等于所述损失阈值时,则表示所述损失值不满足所述预设条件时。其中,所述损失阈值可以设置为0.1,也可以根据实际场景设置。进一步地,所述语音健康识别模型的参数调整可以通过当前已知的随机梯度下降算法实现,在此不做进一步赘述。In an optional embodiment of the present application, the preset condition includes that the loss value is less than a loss threshold. That is, when the loss value is less than the loss threshold, it means that the loss value satisfies the preset condition, and when the loss value is greater than or equal to the loss threshold, it means that the loss value does not meet the predetermined condition. when the preset conditions are stated. Wherein, the loss threshold may be set to 0.1, or may be set according to actual scenarios. Further, the parameter adjustment of the voice health recognition model may be implemented by a currently known stochastic gradient descent algorithm, which will not be described further herein.
S3、利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果。S3. Use the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result.
本申请实施例中,所述舌相健康识别模型包括图像分类模块和图像分析模块,其中,所述图像分类模块用于对所述舌相图像进行背景分割,以输出舌相区域,所述图像分析模块用于对所述图像分类模块输出的舌相区域进行舌相健康分析,输出舌相健康状态。In the embodiment of the present application, the tongue health recognition model includes an image classification module and an image analysis module, wherein the image classification module is used to perform background segmentation on the tongue image to output the tongue region, the image The analysis module is configured to perform a tongue health analysis on the tongue region output by the image classification module, and output the tongue health state.
本申请实施例中,所述利用训练完成的舌相健康识别模型对所述舌相图像进行舌相健康检测之前,需要对所述舌相健康识别模型进行训练,以保障所述舌相健康识别模型的舌相健康检测准确率。In the embodiment of the present application, before the tongue health detection is performed on the tongue image by using the trained tongue health recognition model, the tongue health recognition model needs to be trained to ensure the tongue health recognition. The model's tongue health detection accuracy.
详细地,所述对所述舌相健康识别模型进行训练,包括:获取训练舌相图像,利用所述舌相健康识别模型中的图像分类模块对所述训练舌相图像进行特征提取,得到特征舌相图像,利用所述舌相健康识别模型中的图像分析模块检测所述特征舌相图像的健康状态,得到预测健康状态,计算所述预测健康状态与对应所述训练舌相图像的标准健康状态的训练损失,根据所述训练损失,调整所述舌相健康识别模型的参数,直至所述训练损失小于预设训练损失时,得到训练完成的舌相健康识别模型。可选的,所述预设训练损失为0.1。In detail, the training of the tongue health recognition model includes: acquiring a training tongue image, and using an image classification module in the tongue health recognition model to perform feature extraction on the training tongue image to obtain features Tongue image, use the image analysis module in the tongue health recognition model to detect the health status of the characteristic tongue image, obtain the predicted health status, and calculate the predicted health status and the standard health corresponding to the training tongue image The training loss of the state, according to the training loss, the parameters of the tongue health recognition model are adjusted until the training loss is less than the preset training loss, and the trained tongue health recognition model is obtained. Optionally, the preset training loss is 0.1.
在本申请的一个可选实施例中,所述利用所述舌相健康识别模型中的图像分类模块对所述训练舌相图像进行特征提取,得到特征舌相图像,包括:利用所述图像分类模块中的卷积层对所述训练舌相图像进行卷积操作,得到初始特征舌相图像,利用所述图像分类模块中的池化层对所述初始特征舌相图像进行降维操作,得到降维特征舌相图像,利用所述图像分类模块中的激活函数输出所述降维特征舌相图像,得到所述特征舌相图像。其中,所述图像分类模块中的激活函数包括relu激活函数。In an optional embodiment of the present application, performing feature extraction on the training lingual image by using the image classification module in the lingual health recognition model to obtain a characteristic lingual image includes: using the image classification The convolution layer in the module performs the convolution operation on the training tongue image to obtain the initial characteristic tongue image, and uses the pooling layer in the image classification module to perform a dimensionality reduction operation on the initial characteristic tongue image to obtain The dimension-reduced characteristic lingual image is outputted by using the activation function in the image classification module to obtain the characteristic lingual image. Wherein, the activation function in the image classification module includes a relu activation function.
在本申请的一个可选实施例中,所述利用所述舌相健康识别模型中的图像分析模块检测所述特征舌相图像的健康状态,得到预测健康状态,包括:利用所述图像分析模块中的采样层对所述特征舌相图像进行上采样,得到采样舌相图像,利用所述图像分析模块中的全连接层对所述采样舌相图像进行健康检测后并输出,得到所述预测健康状态。In an optional embodiment of the present application, the detecting the health state of the characteristic tongue image by using the image analysis module in the tongue health recognition model to obtain the predicted health state includes: using the image analysis module The sampling layer in the sample up-samples the characteristic lingual image to obtain the sampled lingual image, and uses the fully connected layer in the image analysis module to perform health detection on the sampled lingual image and output it to obtain the prediction. health status.
在本申请的一个可选实施例中,所述训练损失可以通过当前已知的sigmoid函数计算。In an optional embodiment of the present application, the training loss may be calculated by a currently known sigmoid function.
进一步地,本申请实施例中,所述利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测之前,还包括:对所述舌相图像进行预处理操作,以改善舌相图像的质量,保证舌相图像的分析准确性。其中,所述预处理操作包括:通过各比例法将所述舌相图像执行灰度转换操作,得到灰度舌相图像;利用高斯滤波对所述灰度舌相图像进行减噪; 并利用对比度增强对减噪后的所述灰度舌相图像进行对比度增强;根据OTSU算法将对比度增强后的所述灰度舌相图像进行阈值化操作。Further, in the embodiment of the present application, before performing image health detection on the lingual image by using the trained lingual health recognition model, the method further includes: performing a preprocessing operation on the lingual image to improve the lingual image. The quality of the images ensures the accuracy of the analysis of lingual images. Wherein, the preprocessing operation includes: performing a grayscale conversion operation on the tongue image through each proportional method to obtain a grayscale tongue image; using Gaussian filtering to reduce noise on the grayscale tongue image; and using contrast The enhancement is to perform contrast enhancement on the noise-reduced grayscale lingual image; and perform a thresholding operation on the contrast-enhanced grayscale lingual image according to the OTSU algorithm.
S6、对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。S6. Integrate the voice health detection result and the tongue health detection result to obtain a health detection result, and push the health detection result to the user.
本申请实施例中,利用下述公式对所述语音健康检测结果和所述舌相健康检测结果进行融合:In the embodiment of the present application, the following formula is used to fuse the voice health detection result and the tongue health detection result:
Figure PCTCN2021082864-appb-000004
Figure PCTCN2021082864-appb-000004
其中,f(x,a)表示健康检测结果,k表示融合的所述健康检测结果的数量,x表示所述语音健康检测结果和所述舌相健康检测结果的相同矢量,
Figure PCTCN2021082864-appb-000005
表示所述语音健康检测结果,
Figure PCTCN2021082864-appb-000006
表示所述舌相健康检测结果,ɑ表示预设的权重参数(a∈(0,1))。
Wherein, f(x, a) represents the health detection result, k represents the number of the fused health detection results, x represents the same vector of the voice health detection result and the tongue health detection result,
Figure PCTCN2021082864-appb-000005
represents the voice health detection result,
Figure PCTCN2021082864-appb-000006
represents the tongue health detection result, and ɑ represents the preset weight parameter (a∈(0,1)).
进一步地,需要说明的是,所述权重参数的划分可以根据实际业务场景实现,比如划分舌相健康检测结果的权重为60%,所述语音健康检测结果的权重为40%。Further, it should be noted that the division of the weight parameters can be implemented according to actual business scenarios, for example, the weight of the tongue health detection result is 60%, and the weight of the voice health detection result is 40%.
进一步地,本申请实施例可以通过移动端将所述健康检测结果推送至所述用户,以方便用户可以实时直观的了解自身身体状态,其中,所述移动端可以为手机端。Further, in this embodiment of the present application, the health detection result can be pushed to the user through a mobile terminal, so that the user can intuitively understand his physical state in real time, wherein the mobile terminal can be a mobile terminal.
进一步地,为保障所述健康检测结果的隐私性和安全性,所述健康检测结果还可存储于一区块链节点中。Further, in order to ensure the privacy and security of the health detection results, the health detection results can also be stored in a blockchain node.
本申请实施例首先获取用户的语音数据和舌相图像,以保障后续用户健康检测的前提;其次,本申请实施例利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果,并利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果,可以实现线上用户的语音数据和舌相数据的健康检测,使得用户的健康检测更加全面;进一步地,本申请实施例对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户,以帮助用户实时直观的了解到自身身体状态。因此,本申请提可以提高用户健康检测的便利性。The embodiment of the present application first obtains the user's voice data and tongue image to ensure the premise of subsequent user health detection; secondly, the embodiment of the present application uses the trained voice health recognition model to perform voice health detection on the voice data to obtain a voice Health detection results, and use the trained tongue health recognition model to perform image health detection on the tongue images to obtain tongue health detection results, which can realize the health detection of online users' voice data and tongue data, so that users can The health detection is more comprehensive; further, the embodiment of the present application fuses the voice health detection result and the tongue health detection result to obtain a health detection result, and pushes the health detection result to the user to Help users understand their physical state intuitively in real time. Therefore, the present application can improve the convenience of user health detection.
如图3所示,是本申请健康检测装置的功能模块图。As shown in FIG. 3 , it is a functional block diagram of the health detection device of the present application.
本申请所述健康检测装置100可以安装于电子设备中。根据实现的功能,所述健康检测装置可以包括获取模块101、检测模块102以及推送模块103。本发所述模块也可以称之为单元,是指一种能够被电子设备处理器所执行,并且能够完成固定功能的一系列计算机程序段,其存储在电子设备的存储器中。The health detection apparatus 100 described in this application may be installed in an electronic device. According to the implemented functions, the health detection apparatus may include an acquisition module 101 , a detection module 102 and a push module 103 . The modules described in the present invention can also be called units, which refer to a series of computer program segments that can be executed by the electronic device processor and can perform fixed functions, and are stored in the memory of the electronic device.
在本实施例中,关于各模块/单元的功能如下:In this embodiment, the functions of each module/unit are as follows:
所述获取模块101,用于获取用户的语音数据和舌相图像;The acquisition module 101 is used to acquire the user's voice data and tongue image;
所述检测模块102,用于利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;The detection module 102 is configured to perform voice health detection on the voice data by using the trained voice health recognition model to obtain a voice health detection result;
所述检测模块102,用于利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;The detection module 102 is configured to perform image health detection on the tongue image by using the trained tongue health recognition model to obtain a tongue health detection result;
所述推送模块103,用于对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The pushing module 103 is configured to fuse the voice health detection result and the tongue health detection result to obtain a health detection result, and push the health detection result to the user.
详细地,本申请实施例中所述健康检测装置100中的所述各模块在使用时采用与上述的图1和图2中所述的健康检测方法一样的技术手段,并能够产生相同的技术效果,这里不再赘述。In detail, the modules in the health detection device 100 in the embodiments of the present application use the same technical means as the health detection methods described in the above-mentioned FIG. 1 and FIG. 2 , and can generate the same technology The effect will not be repeated here.
如图4所示,是本申请实现健康检测方法的电子设备的结构示意图。As shown in FIG. 4 , it is a schematic structural diagram of an electronic device implementing the health detection method of the present application.
所述电子设备1可以包括处理器10、存储器11和总线,还可以包括存储在所述存储器11中并可在所述处理器10上运行的计算机程序,如健康检测程序12。The electronic device 1 may include a processor 10, a memory 11 and a bus, and may also include a computer program stored in the memory 11 and executable on the processor 10, such as a health detection program 12.
其中,所述存储器11至少包括一种类型的可读存储介质,所述可读存储介质包括闪存、移动硬盘、多媒体卡、卡型存储器(例如:SD或DX存储器等)、磁性存储器、磁盘、光 盘等。所述存储器11在一些实施例中可以是电子设备1的内部存储单元,例如该电子设备1的移动硬盘。所述存储器11在另一些实施例中也可以是电子设备1的外部存储设备,例如电子设备1上配备的插接式移动硬盘、智能存储卡(Smart Media Card,SMC)、安全数字(Secure Digital,SD)卡、闪存卡(Flash Card)等。进一步地,所述存储器11还可以既包括电子设备1的内部存储单元也包括外部存储设备。所述存储器11不仅可以用于存储安装于电子设备1的应用软件及各类数据,例如健康检测程序12的代码等,还可以用于暂时地存储已经输出或者将要输出的数据。Wherein, the memory 11 includes at least one type of readable storage medium, and the readable storage medium includes flash memory, mobile hard disk, multimedia card, card-type memory (for example: SD or DX memory, etc.), magnetic memory, magnetic disk, CD etc. The memory 11 may be an internal storage unit of the electronic device 1 in some embodiments, such as a mobile hard disk of the electronic device 1 . In other embodiments, the memory 11 may also be an external storage device of the electronic device 1, such as a pluggable mobile hard disk, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital) equipped on the electronic device 1. , SD) card, flash memory card (Flash Card), etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device. The memory 11 can not only be used to store application software installed in the electronic device 1 and various types of data, such as codes of the health detection program 12, etc., but also can be used to temporarily store data that has been output or will be output.
所述处理器10在一些实施例中可以由集成电路组成,例如可以由单个封装的集成电路所组成,也可以是由多个相同功能或不同功能封装的集成电路所组成,包括一个或者多个中央处理器(Central Processing unit,CPU)、微处理器、数字处理芯片、图形处理器及各种控制芯片的组合等。所述处理器10是所述电子设备的控制核心(Control Unit),利用各种接口和线路连接整个电子设备的各个部件,通过运行或执行存储在所述存储器11内的程序或者模块(例如执行健康检测程序等),以及调用存储在所述存储器11内的数据,以执行电子设备1的各种功能和处理数据。In some embodiments, the processor 10 may be composed of integrated circuits, for example, may be composed of a single packaged integrated circuit, or may be composed of multiple integrated circuits packaged with the same function or different functions, including one or more integrated circuits. Central Processing Unit (CPU), microprocessor, digital processing chip, graphics processor and combination of various control chips, etc. The processor 10 is the control core (Control Unit) of the electronic device, and uses various interfaces and lines to connect the various components of the entire electronic device, by running or executing the program or module (for example, executing the program) stored in the memory 11. health detection program, etc.), and call the data stored in the memory 11 to perform various functions of the electronic device 1 and process data.
所述总线可以是外设部件互连标准(peripheral component interconnect,简称PCI)总线或扩展工业标准结构(extended industry standard architecture,简称EISA)总线等。该总线可以分为地址总线、数据总线、控制总线等。所述总线被设置为实现所述存储器11以及至少一个处理器10等之间的连接通信。The bus may be a peripheral component interconnect (PCI for short) bus or an extended industry standard architecture (Extended industry standard architecture, EISA for short) bus or the like. The bus can be divided into address bus, data bus, control bus and so on. The bus is configured to implement connection communication between the memory 11 and at least one processor 10 and the like.
图4仅示出了具有部件的电子设备,本领域技术人员可以理解的是,图4示出的结构并不构成对所述电子设备1的限定,可以包括比图示更少或者更多的部件,或者组合某些部件,或者不同的部件布置。FIG. 4 only shows an electronic device with components. Those skilled in the art can understand that the structure shown in FIG. 4 does not constitute a limitation on the electronic device 1, and may include fewer or more components than those shown in the drawings. components, or a combination of certain components, or a different arrangement of components.
例如,尽管未示出,所述电子设备1还可以包括给各个部件供电的电源(比如电池),优选地,电源可以通过电源管理装置与所述至少一个处理器10逻辑相连,从而通过电源管理装置实现充电管理、放电管理、以及功耗管理等功能。电源还可以包括一个或一个以上的直流或交流电源、再充电装置、电源故障检测电路、电源转换器或者逆变器、电源状态指示器等任意组件。所述电子设备1还可以包括多种传感器、蓝牙模块、Wi-Fi模块等,在此不再赘述。For example, although not shown, the electronic device 1 may also include a power supply (such as a battery) for powering the various components, preferably, the power supply may be logically connected to the at least one processor 10 through a power management device, so that the power management The device implements functions such as charge management, discharge management, and power consumption management. The power source may also include one or more DC or AC power sources, recharging devices, power failure detection circuits, power converters or inverters, power status indicators, and any other components. The electronic device 1 may further include various sensors, Bluetooth modules, Wi-Fi modules, etc., which will not be repeated here.
进一步地,所述电子设备1还可以包括网络接口,可选地,所述网络接口可以包括有线接口和/或无线接口(如WI-FI接口、蓝牙接口等),通常用于在该电子设备1与其他电子设备之间建立通信连接。Further, the electronic device 1 may also include a network interface, optionally, the network interface may include a wired interface and/or a wireless interface (such as a WI-FI interface, a Bluetooth interface, etc.), which is usually used in the electronic device 1 Establish a communication connection with other electronic devices.
可选地,该电子设备1还可以包括用户接口,用户接口可以是显示器(Display)、输入单元(比如键盘(Keyboard)),可选地,用户接口还可以是标准的有线接口、无线接口。可选地,在一些实施例中,显示器可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。其中,显示器也可以适当的称为显示屏或显示单元,用于显示在电子设备1中处理的信息以及用于显示可视化的用户界面。Optionally, the electronic device 1 may further include a user interface, and the user interface may be a display (Display), an input unit (eg, a keyboard (Keyboard)), optionally, the user interface may also be a standard wired interface or a wireless interface. Optionally, in some embodiments, the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light-emitting diode) touch device, and the like. The display may also be appropriately called a display screen or a display unit, which is used for displaying information processed in the electronic device 1 and for displaying a visualized user interface.
应该了解,所述实施例仅为说明之用,在专利申请范围上并不受此结构的限制。It should be understood that the embodiments are only used for illustration, and are not limited by this structure in the scope of the patent application.
所述电子设备1中的所述存储器11存储的健康检测程序12是多个程序的组合,在所述处理器10中运行时,可以实现:The health detection program 12 stored in the memory 11 in the electronic device 1 is a combination of multiple programs. When running in the processor 10, it can realize:
获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并 将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
具体地,所述处理器10对上述程序的具体实现方法可参考图1及图2对应实施例中相关步骤的描述,在此不赘述。Specifically, for the specific implementation method of the above program by the processor 10, reference may be made to the description of the relevant steps in the corresponding embodiments of FIG. 1 and FIG. 2 , and details are not described herein.
进一步地,所述电子设备1集成的模块/单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个非易失性计算机可读取存储介质中。所述计算机可读存储介质可以是易失性的,也可以是非易失性的。例如,所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)。Further, if the modules/units integrated in the electronic device 1 are implemented in the form of software functional units and sold or used as independent products, they may be stored in a non-volatile computer-readable storage medium. The computer-readable storage medium may be volatile or non-volatile. For example, the computer-readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a removable hard disk, a magnetic disk, an optical disc, a computer memory, a read-only memory (ROM, Read-Only). Memory).
本申请还提供一种计算机可读存储介质,所述可读存储介质存储有计算机程序,所述计算机程序在被电子设备的处理器所执行时,可以实现:The present application also provides a computer-readable storage medium, where the readable storage medium stores a computer program, and when executed by a processor of an electronic device, the computer program can realize:
获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
在本申请所提供的几个实施例中,应该理解到,所揭露的设备,装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述模块的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式。In the several embodiments provided in this application, it should be understood that the disclosed apparatus, apparatus and method may be implemented in other manners. For example, the apparatus embodiments described above are only illustrative. For example, the division of the modules is only a logical function division, and there may be other division manners in actual implementation.
所述作为分离部件说明的模块可以是或者也可以不是物理上分开的,作为模块显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。The modules described as separate components may or may not be physically separated, and components shown as modules may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
另外,在本申请各个实施例中的各功能模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能模块的形式实现。In addition, each functional module in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit. The above-mentioned integrated units can be implemented in the form of hardware, or can be implemented in the form of hardware plus software function modules.
对于本领域技术人员而言,显然本申请不限于上述示范性实施例的细节,而且在不背离本申请的精神或基本特征的情况下,能够以其他的具体形式实现本申请。It will be apparent to those skilled in the art that the present application is not limited to the details of the above-described exemplary embodiments, but that the present application can be implemented in other specific forms without departing from the spirit or essential characteristics of the present application.
因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本申请的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化涵括在本申请内。不应将权利要求中的任何附关联图标记视为限制所涉及的权利要求。Accordingly, the embodiments are to be regarded in all respects as illustrative and not restrictive, and the scope of the application is to be defined by the appended claims rather than the foregoing description, which is therefore intended to fall within the scope of the claims. All changes within the meaning and scope of the equivalents of , are included in this application. Any reference signs in the claims shall not be construed as limiting the involved claim.
本申请所指区块链是分布式数据存储、点对点传输、共识机制、加密算法等计算机技术的新型应用模式。区块链(Blockchain),本质上是一个去中心化的数据库,是一串使用密码学方法相关联产生的数据块,每一个数据块中包含了一批次网络交易的信息,用于验证其信息的有效性(防伪)和生成下一个区块。区块链可以包括区块链底层平台、平台产品服务层以及应用服务层等。The blockchain referred to in this application is a new application mode of computer technologies such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm. Blockchain, essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information to verify its Validity of information (anti-counterfeiting) and generation of the next block. The blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
此外,显然“包括”一词不排除其他单元或步骤,单数不排除复数。系统权利要求中陈述的多个单元或装置也可以由一个单元或装置通过软件或者硬件来实现。第二等词语用来表示名称,而并不表示任何特定的顺序。Furthermore, it is clear that the word "comprising" does not exclude other units or steps and the singular does not exclude the plural. Several units or means recited in the system claims can also be realized by one unit or means by means of software or hardware. Second-class terms are used to denote names and do not denote any particular order.
最后应说明的是,以上实施例仅用以说明本申请的技术方案而非限制,尽管参照较佳实施例对本申请进行了详细说明,本领域的普通技术人员应当理解,可以对本申请的技术方案进行修改或等同替换,而不脱离本申请技术方案的精神和范围。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present application rather than limitations. Although the present application has been described in detail with reference to the preferred embodiments, those of ordinary skill in the art should understand that the technical solutions of the present application can be Modifications or equivalent substitutions can be made without departing from the spirit and scope of the technical solutions of the present application.

Claims (20)

  1. 一种健康检测方法,其中,所述方法包括:A health detection method, wherein the method includes:
    获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
    利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
    利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
    对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  2. 如权利要求1所述的健康检测方法,其中,所述利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测之前,所述方法还包括:The health detection method according to claim 1, wherein before the voice health detection is performed on the voice data by using the trained voice health recognition model, the method further comprises:
    获取训练语音数据,标记所述训练语音数据中的人声数据,得到标准人声数据,并对所述标准人声数据进行健康检测,得到标准人声健康数据;Obtaining training voice data, marking the human voice data in the training voice data, obtaining standard human voice data, and performing health detection on the standard human voice data to obtain standard human voice health data;
    利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据;Use the voice classification module of the voice health recognition model to perform human voice segmentation on the training voice data to obtain training voice data;
    利用所述语音健康识别模型中的语音分析模块对所述训练人声数据进行健康检测,得到训练人声健康数据;Use the voice analysis module in the voice health recognition model to perform health detection on the training vocal data to obtain training vocal health data;
    根据所述标准人声数据、所述标准人声健康数据、所述训练人声数据以及所述训练人声健康数据,计算所述语音健康识别模型的损失值;Calculate the loss value of the voice health recognition model according to the standard vocal data, the standard vocal health data, the training vocal data, and the training vocal health data;
    若所述损失值不满足预设条件,则调整所述语音健康识别模型的参数,并返回所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据的步骤;If the loss value does not meet the preset conditions, then adjust the parameters of the voice health recognition model, and return to the voice classification module using the voice health recognition model to perform human voice segmentation on the training voice data to obtain training Steps for vocal data;
    若所述损失值满足预设条件,则得到训练完成的语音健康识别模型。If the loss value satisfies the preset condition, a trained voice health recognition model is obtained.
  3. 如权利要求2所述的健康检测方法,其中,所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据,包括:The health detection method according to claim 2, wherein the voice classification module using the voice health recognition model performs human voice segmentation on the training voice data to obtain training voice data, comprising:
    利用所述语音分类模块中的声音频率转换算法将所述训练语音数据转换成对应的语音频率,并计算所述语音频率的维度参数;Using the voice frequency conversion algorithm in the voice classification module to convert the training voice data into corresponding voice frequencies, and calculate the dimension parameters of the voice frequencies;
    根据所述维度参数,筛选出所述训练语音数据中的人声数据,得到训练人声数据。According to the dimension parameter, the human voice data in the training voice data is screened out to obtain training human voice data.
  4. 如权利要求2所述的健康检测方法,其中,所述根据所述标准人声数据、标准人声健康数据、训练人声数据以及所述训练人声健康数据,计算所述语音健康识别模型的损失值,包括:The health detection method according to claim 2, wherein the calculation of the voice health recognition model is based on the standard vocal data, standard vocal health data, training vocal data and the training vocal health data. Loss values, including:
    根据所述标准人声数据和所述训练人声数据,计算所述语音健康识别模型的第一损失值;Calculate the first loss value of the voice health recognition model according to the standard vocal data and the training vocal data;
    根据所述标准人声健康数据和所述训练人声健康数据,计算所述语音健康识别模型的第二损失值;Calculate the second loss value of the voice health recognition model according to the standard vocal health data and the training vocal health data;
    根据所述第一损失值和所述第二损失值,计算所述语音健康识别模型的损失值。According to the first loss value and the second loss value, a loss value of the speech health recognition model is calculated.
  5. 如权利要求1中所述的健康检测方法,其中,所述利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测之前,所述方法还包括:The health detection method according to claim 1, wherein, before performing image health detection on the tongue image by using the trained tongue health recognition model, the method further comprises:
    获取训练舌相图像;Obtain training tongue images;
    利用所述舌相健康识别模型中的图像分类模块对所述训练舌相图像进行特征提取,得到特征舌相图像;Using the image classification module in the lingual health recognition model to perform feature extraction on the training lingual images to obtain characteristic lingual images;
    利用所述舌相健康识别模型中的图像分析模块检测所述特征舌相图像的健康状态,得到预测健康状态;Use the image analysis module in the tongue health recognition model to detect the health state of the characteristic tongue image to obtain the predicted health state;
    计算所述预测健康状态与对应所述训练舌相图像的标准健康状态的训练损失;calculating the training loss of the predicted health state and the standard health state corresponding to the training lingual image;
    根据所述训练损失,调整所述舌相健康识别模型的参数,直至所述训练损失小于预设训练损失时,得到训练完成的舌相健康识别模型。According to the training loss, the parameters of the tongue health recognition model are adjusted until the training loss is less than the preset training loss, and the trained tongue health recognition model is obtained.
  6. 如权利要求1所述的健康检测方法,其中,所述利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测之前,所述方法还包括:The health detection method according to claim 1, wherein, before performing the image health detection on the tongue image by using the trained tongue health recognition model, the method further comprises:
    将所述舌相图像执行灰度转换操作,得到灰度舌相图像;performing a grayscale conversion operation on the lingual image to obtain a grayscale lingual image;
    对所述灰度舌相图像进行减噪;denoising the grayscale lingual image;
    对减噪后的所述灰度舌相图像进行对比度增强;performing contrast enhancement on the grayscale tongue phase image after noise reduction;
    将对比度增强后的所述灰度舌相图像进行阈值化操作。Thresholding is performed on the contrast-enhanced grayscale lingual image.
  7. 如权利要求1至6中任一项所述的健康检测方法,其中,所述对所述语音健康检测结果和所述舌相健康检测结果进行融合,包括:The health detection method according to any one of claims 1 to 6, wherein the fusion of the voice health detection result and the tongue health detection result comprises:
    利用下述公式对所述语音健康检测结果和所述舌相健康检测结果进行融合:Utilize the following formula to fuse the voice health detection result and the tongue health detection result:
    Figure PCTCN2021082864-appb-100001
    Figure PCTCN2021082864-appb-100001
    其中,f(x,a)表示健康检测结果,k表示融合的所述健康检测结果的数量,x表示所述语音健康检测结果和所述舌相健康检测结果的相同矢量,
    Figure PCTCN2021082864-appb-100002
    表示所述语音健康检测结果,
    Figure PCTCN2021082864-appb-100003
    表示所述舌相健康检测结果,ɑ表示预设的权重参数。
    Wherein, f(x, a) represents the health detection result, k represents the number of the fused health detection results, x represents the same vector of the voice health detection result and the tongue health detection result,
    Figure PCTCN2021082864-appb-100002
    represents the voice health detection result,
    Figure PCTCN2021082864-appb-100003
    represents the tongue health detection result, and ɑ represents the preset weight parameter.
  8. 一种健康检测装置,其中,所述装置包括:A health detection device, wherein the device includes:
    获取模块,用于获取用户的语音数据和舌相图像;The acquisition module is used to acquire the user's voice data and tongue image;
    检测模块,用于利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;a detection module, configured to perform voice health detection on the voice data by using the trained voice health recognition model to obtain a voice health detection result;
    所述检测模块,用于利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;The detection module is used to perform image health detection on the tongue image by using the trained tongue health recognition model to obtain a tongue health detection result;
    推送模块,用于对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。A push module is configured to fuse the voice health detection result and the tongue health detection result to obtain a health detection result, and push the health detection result to the user.
  9. 一种电子设备,其中,所述电子设备包括:An electronic device, wherein the electronic device comprises:
    至少一个处理器;以及,at least one processor; and,
    与所述至少一个处理器通信连接的存储器;其中,a memory communicatively coupled to the at least one processor; wherein,
    所述存储器存储有可被所述至少一个处理器执行的计算机程序,所述计算机程序被所述至少一个处理器执行,以使所述至少一个处理器能够执行如下步骤:The memory stores a computer program executable by the at least one processor, the computer program being executed by the at least one processor to enable the at least one processor to perform the steps of:
    获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
    利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
    利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
    对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  10. 如权利要求9所述的电子设备,其中,所述利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测之前,所述计算机程序被所述至少一个处理器执行时还实现如下步骤:The electronic device according to claim 9, wherein before the voice health detection is performed on the voice data by using the trained voice health recognition model, the computer program further implements the following steps when executed by the at least one processor :
    获取训练语音数据,标记所述训练语音数据中的人声数据,得到标准人声数据,并对所述标准人声数据进行健康检测,得到标准人声健康数据;Obtaining training voice data, marking the human voice data in the training voice data, obtaining standard human voice data, and performing health detection on the standard human voice data to obtain standard human voice health data;
    利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据;Use the voice classification module of the voice health recognition model to perform human voice segmentation on the training voice data to obtain training voice data;
    利用所述语音健康识别模型中的语音分析模块对所述训练人声数据进行健康检测,得到训练人声健康数据;Use the voice analysis module in the voice health recognition model to perform health detection on the training vocal data to obtain training vocal health data;
    根据所述标准人声数据、所述标准人声健康数据、所述训练人声数据以及所述训练人 声健康数据,计算所述语音健康识别模型的损失值;Calculate the loss value of the voice health recognition model according to the standard vocal data, the standard vocal health data, the training vocal data and the training vocal health data;
    若所述损失值不满足预设条件,则调整所述语音健康识别模型的参数,并返回所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据的步骤;If the loss value does not meet the preset conditions, adjust the parameters of the voice health recognition model, and return to the voice classification module using the voice health recognition model to perform human voice segmentation on the training voice data to obtain training Steps for vocal data;
    若所述损失值满足预设条件,则得到训练完成的语音健康识别模型。If the loss value satisfies the preset condition, a trained voice health recognition model is obtained.
  11. 如权利要求10所述的电子设备,其中,所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据,包括:The electronic device according to claim 10, wherein the voice classification module using the voice health recognition model performs human voice segmentation on the training voice data to obtain training voice data, comprising:
    利用所述语音分类模块中的声音频率转换算法将所述训练语音数据转换成对应的语音频率,并计算所述语音频率的维度参数;Using the voice frequency conversion algorithm in the voice classification module to convert the training voice data into corresponding voice frequencies, and calculate the dimension parameters of the voice frequencies;
    根据所述维度参数,筛选出所述训练语音数据中的人声数据,得到训练人声数据。According to the dimension parameter, the human voice data in the training voice data is screened out to obtain training human voice data.
  12. 如权利要求10所述的电子设备,其中,所述根据所述标准人声数据、标准人声健康数据、训练人声数据以及所述训练人声健康数据,计算所述语音健康识别模型的损失值,包括:The electronic device of claim 10, wherein the loss of the speech health recognition model is calculated according to the standard vocal data, standard vocal health data, training vocal data, and the training vocal health data values, including:
    根据所述标准人声数据和所述训练人声数据,计算所述语音健康识别模型的第一损失值;Calculate the first loss value of the voice health recognition model according to the standard vocal data and the training vocal data;
    根据所述标准人声健康数据和所述训练人声健康数据,计算所述语音健康识别模型的第二损失值;Calculate the second loss value of the voice health recognition model according to the standard vocal health data and the training vocal health data;
    根据所述第一损失值和所述第二损失值,计算所述语音健康识别模型的损失值。According to the first loss value and the second loss value, a loss value of the speech health recognition model is calculated.
  13. 如权利要求9中所述的电子设备,其中,所述利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测之前,所述计算机程序被所述至少一个处理器执行时还实现如下步骤:The electronic device as claimed in claim 9, wherein before the image health detection is performed on the tongue image by using the trained tongue health recognition model, the computer program is further executed by the at least one processor. Implement the following steps:
    获取训练舌相图像;Obtain training tongue images;
    利用所述舌相健康识别模型中的图像分类模块对所述训练舌相图像进行特征提取,得到特征舌相图像;Using the image classification module in the lingual health recognition model to perform feature extraction on the training lingual images to obtain characteristic lingual images;
    利用所述舌相健康识别模型中的图像分析模块检测所述特征舌相图像的健康状态,得到预测健康状态;Use the image analysis module in the tongue health recognition model to detect the health state of the characteristic tongue image to obtain the predicted health state;
    计算所述预测健康状态与对应所述训练舌相图像的标准健康状态的训练损失;calculating the training loss of the predicted health state and the standard health state corresponding to the training lingual image;
    根据所述训练损失,调整所述舌相健康识别模型的参数,直至所述训练损失小于预设训练损失时,得到训练完成的舌相健康识别模型。According to the training loss, the parameters of the tongue health recognition model are adjusted until the training loss is less than the preset training loss, and the trained tongue health recognition model is obtained.
  14. 如权利要求9所述的电子设备,其中,所述利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测之前,所述计算机程序被所述至少一个处理器执行时还实现如下步骤:The electronic device according to claim 9, wherein before the image health detection is performed on the tongue image by using the trained tongue health recognition model, the computer program is further implemented when executed by the at least one processor Follow the steps below:
    将所述舌相图像执行灰度转换操作,得到灰度舌相图像;performing a grayscale conversion operation on the lingual image to obtain a grayscale lingual image;
    对所述灰度舌相图像进行减噪;denoising the grayscale lingual image;
    对减噪后的所述灰度舌相图像进行对比度增强;performing contrast enhancement on the grayscale tongue phase image after noise reduction;
    将对比度增强后的所述灰度舌相图像进行阈值化操作。Thresholding is performed on the contrast-enhanced grayscale lingual image.
  15. 如权利要求9至14中任一项所述的电子设备,其中,所述对所述语音健康检测结果和所述舌相健康检测结果进行融合,包括:The electronic device according to any one of claims 9 to 14, wherein the fusion of the voice health detection result and the tongue health detection result comprises:
    利用下述公式对所述语音健康检测结果和所述舌相健康检测结果进行融合:Utilize the following formula to fuse the voice health detection result and the tongue health detection result:
    Figure PCTCN2021082864-appb-100004
    Figure PCTCN2021082864-appb-100004
    其中,f(x,a)表示健康检测结果,k表示融合的所述健康检测结果的数量,x表示所述语音健康检测结果和所述舌相健康检测结果的相同矢量,
    Figure PCTCN2021082864-appb-100005
    表示所述语音健康检测结果,
    Figure PCTCN2021082864-appb-100006
    表示所述舌相健康检测结果,ɑ表示预设的权重参数。
    Wherein, f(x, a) represents the health detection result, k represents the number of the fused health detection results, x represents the same vector of the voice health detection result and the tongue health detection result,
    Figure PCTCN2021082864-appb-100005
    represents the voice health detection result,
    Figure PCTCN2021082864-appb-100006
    represents the tongue health detection result, and ɑ represents the preset weight parameter.
  16. 一种计算机可读存储介质,存储有计算机程序,其中,所述计算机程序被处理器执行时实现如下步骤:A computer-readable storage medium storing a computer program, wherein the computer program implements the following steps when executed by a processor:
    获取用户的语音数据和舌相图像;Obtain the user's voice data and tongue images;
    利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测,得到语音健康检测结果;Use the trained voice health recognition model to perform voice health detection on the voice data, and obtain a voice health detection result;
    利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测,得到舌相健康检测结果;Using the trained tongue health recognition model to perform image health detection on the tongue image to obtain a tongue health detection result;
    对所述语音健康检测结果和所述舌相健康检测结果进行融合,得到健康检测结果,并将所述健康检测结果推送至所述用户。The voice health detection result and the tongue health detection result are fused to obtain a health detection result, and the health detection result is pushed to the user.
  17. 如权利要求16所述的计算机可读存储介质,其中,所述利用训练完成的语音健康识别模型对所述语音数据进行语音健康检测之前,所述计算机程序被处理器执行时还实现如下步骤:The computer-readable storage medium according to claim 16, wherein, before the voice health detection is performed on the voice data by using the trained voice health recognition model, the computer program further implements the following steps when executed by the processor:
    获取训练语音数据,标记所述训练语音数据中的人声数据,得到标准人声数据,并对所述标准人声数据进行健康检测,得到标准人声健康数据;Obtaining training voice data, marking the human voice data in the training voice data, obtaining standard human voice data, and performing health detection on the standard human voice data to obtain standard human voice health data;
    利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据;Use the voice classification module of the voice health recognition model to perform human voice segmentation on the training voice data to obtain training voice data;
    利用所述语音健康识别模型中的语音分析模块对所述训练人声数据进行健康检测,得到训练人声健康数据;Use the voice analysis module in the voice health recognition model to perform health detection on the training vocal data to obtain training vocal health data;
    根据所述标准人声数据、所述标准人声健康数据、所述训练人声数据以及所述训练人声健康数据,计算所述语音健康识别模型的损失值;Calculate the loss value of the voice health recognition model according to the standard vocal data, the standard vocal health data, the training vocal data, and the training vocal health data;
    若所述损失值不满足预设条件,则调整所述语音健康识别模型的参数,并返回所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据的步骤;If the loss value does not meet the preset conditions, then adjust the parameters of the voice health recognition model, and return to the voice classification module using the voice health recognition model to perform human voice segmentation on the training voice data to obtain training Steps for vocal data;
    若所述损失值满足预设条件,则得到训练完成的语音健康识别模型。If the loss value satisfies the preset condition, a trained voice health recognition model is obtained.
  18. 如权利要求17所述的计算机可读存储介质,其中,所述利用所述语音健康识别模型的语音分类模块对所述训练语音数据进行人声分割,得到训练人声数据,包括:The computer-readable storage medium according to claim 17 , wherein the voice segmentation of the training voice data by the voice classification module of the voice health recognition model to obtain training voice data, comprising:
    利用所述语音分类模块中的声音频率转换算法将所述训练语音数据转换成对应的语音频率,并计算所述语音频率的维度参数;Using the voice frequency conversion algorithm in the voice classification module to convert the training voice data into corresponding voice frequencies, and calculate the dimension parameters of the voice frequencies;
    根据所述维度参数,筛选出所述训练语音数据中的人声数据,得到训练人声数据。According to the dimension parameter, the human voice data in the training voice data is screened out to obtain training human voice data.
  19. 如权利要求17所述的计算机可读存储介质,其中,所述根据所述标准人声数据、标准人声健康数据、训练人声数据以及所述训练人声健康数据,计算所述语音健康识别模型的损失值,包括:18. The computer-readable storage medium of claim 17, wherein the voice health recognition is calculated based on the standard vocal data, standard vocal health data, training vocal data, and the training vocal health data The loss value of the model, including:
    根据所述标准人声数据和所述训练人声数据,计算所述语音健康识别模型的第一损失值;Calculate the first loss value of the voice health recognition model according to the standard vocal data and the training vocal data;
    根据所述标准人声健康数据和所述训练人声健康数据,计算所述语音健康识别模型的第二损失值;Calculate the second loss value of the voice health recognition model according to the standard vocal health data and the training vocal health data;
    根据所述第一损失值和所述第二损失值,计算所述语音健康识别模型的损失值。According to the first loss value and the second loss value, a loss value of the speech health recognition model is calculated.
  20. 如权利要求16中所述的计算机可读存储介质,其中,所述利用训练完成的舌相健康识别模型对所述舌相图像进行图像健康检测之前,所述计算机程序被处理器执行时还实现如下步骤:The computer-readable storage medium according to claim 16, wherein before the image health detection is performed on the tongue image by using the trained tongue health recognition model, the computer program further implements when executed by the processor Follow the steps below:
    获取训练舌相图像;Obtain training tongue images;
    利用所述舌相健康识别模型中的图像分类模块对所述训练舌相图像进行特征提取,得到特征舌相图像;Using the image classification module in the lingual health recognition model to perform feature extraction on the training lingual images to obtain characteristic lingual images;
    利用所述舌相健康识别模型中的图像分析模块检测所述特征舌相图像的健康状态,得到预测健康状态;Use the image analysis module in the tongue health recognition model to detect the health state of the characteristic tongue image to obtain the predicted health state;
    计算所述预测健康状态与对应所述训练舌相图像的标准健康状态的训练损失;calculating the training loss of the predicted health state and the standard health state corresponding to the training lingual image;
    根据所述训练损失,调整所述舌相健康识别模型的参数,直至所述训练损失小于预设 训练损失时,得到训练完成的舌相健康识别模型。According to the training loss, the parameters of the tongue health recognition model are adjusted until the training loss is less than the preset training loss, and the trained tongue health recognition model is obtained.
PCT/CN2021/082864 2021-02-26 2021-03-25 Health testing method and apparatus, and device and storage medium WO2022178934A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110214173.3A CN112967806A (en) 2021-02-26 2021-02-26 Health detection method, device, equipment and storage medium
CN202110214173.3 2021-02-26

Publications (1)

Publication Number Publication Date
WO2022178934A1 true WO2022178934A1 (en) 2022-09-01

Family

ID=76275668

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/082864 WO2022178934A1 (en) 2021-02-26 2021-03-25 Health testing method and apparatus, and device and storage medium

Country Status (2)

Country Link
CN (1) CN112967806A (en)
WO (1) WO2022178934A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106682427A (en) * 2016-12-29 2017-05-17 平安科技(深圳)有限公司 Personal health condition assessment method and device based position services
CN109431463A (en) * 2018-10-23 2019-03-08 南开大学 Deep learning Chinese medicine intelligence diagnosis and therapy system based on traditional Chinese and western medicine sample labeling
CN110703965A (en) * 2019-10-11 2020-01-17 上海中医药大学 Intelligent traditional Chinese medicine health state identification software and electronic equipment
US20200160512A1 (en) * 2018-11-16 2020-05-21 Boe Technology Group Co., Ltd. Method, client, server and system for detecting tongue image, and tongue imager
CN112382390A (en) * 2020-11-09 2021-02-19 北京沃东天骏信息技术有限公司 Method, system and storage medium for generating health assessment report

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111899878B (en) * 2020-07-30 2023-06-02 平安科技(深圳)有限公司 Old person health detection system, method, computer device and readable storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106682427A (en) * 2016-12-29 2017-05-17 平安科技(深圳)有限公司 Personal health condition assessment method and device based position services
CN109431463A (en) * 2018-10-23 2019-03-08 南开大学 Deep learning Chinese medicine intelligence diagnosis and therapy system based on traditional Chinese and western medicine sample labeling
US20200160512A1 (en) * 2018-11-16 2020-05-21 Boe Technology Group Co., Ltd. Method, client, server and system for detecting tongue image, and tongue imager
CN110703965A (en) * 2019-10-11 2020-01-17 上海中医药大学 Intelligent traditional Chinese medicine health state identification software and electronic equipment
CN112382390A (en) * 2020-11-09 2021-02-19 北京沃东天骏信息技术有限公司 Method, system and storage medium for generating health assessment report

Also Published As

Publication number Publication date
CN112967806A (en) 2021-06-15

Similar Documents

Publication Publication Date Title
CN110619568A (en) Risk assessment report generation method, device, equipment and storage medium
CN111523389A (en) Intelligent emotion recognition method and device, electronic equipment and storage medium
WO2022116420A1 (en) Speech event detection method and apparatus, electronic device, and computer storage medium
CN108197592B (en) Information acquisition method and device
CN112951203B (en) Speech synthesis method, device, electronic equipment and storage medium
WO2022178933A1 (en) Context-based voice sentiment detection method and apparatus, device and storage medium
CN113064994A (en) Conference quality evaluation method, device, equipment and storage medium
CN114582487A (en) Traditional Chinese medicine diagnosis and treatment assisting method and system based on traditional Chinese medicine knowledge graph
CN116578704A (en) Text emotion classification method, device, equipment and computer readable medium
WO2021208700A1 (en) Method and apparatus for speech data selection, electronic device, and storage medium
CN114155832A (en) Speech recognition method, device, equipment and medium based on deep learning
CN117056728A (en) Time sequence generation method, device, equipment and storage medium
CN112466337A (en) Audio data emotion detection method and device, electronic equipment and storage medium
CN113269719A (en) Model training method, image processing method, device, equipment and storage medium
WO2022178934A1 (en) Health testing method and apparatus, and device and storage medium
CN115223214A (en) Identification method of synthetic mouth-shaped face, model acquisition method, device and equipment
CN116450797A (en) Emotion classification method, device, equipment and medium based on multi-modal dialogue
CN114842880A (en) Intelligent customer service voice rhythm adjusting method, device, equipment and storage medium
CN112652325B (en) Remote voice adjustment method based on artificial intelligence and related equipment
US11457109B2 (en) Relationship determination system
CN111275035B (en) Method and system for identifying background information
CN114495292A (en) Identity recognition method, device, equipment and readable storage medium
Feng et al. Research and application of tongue and face diagnosis based on deep learning
CN113808577A (en) Intelligent extraction method and device of voice abstract, electronic equipment and storage medium
CN112911334A (en) Emotion recognition method, device and equipment based on audio and video data and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21927368

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21927368

Country of ref document: EP

Kind code of ref document: A1