WO2021114623A1 - Method, apparatus, computer device, and storage medium for identifying persons having deformed spinal columns - Google Patents

Method, apparatus, computer device, and storage medium for identifying persons having deformed spinal columns Download PDF

Info

Publication number
WO2021114623A1
WO2021114623A1 PCT/CN2020/099253 CN2020099253W WO2021114623A1 WO 2021114623 A1 WO2021114623 A1 WO 2021114623A1 CN 2020099253 W CN2020099253 W CN 2020099253W WO 2021114623 A1 WO2021114623 A1 WO 2021114623A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
spine
recognized
feature vector
recognition model
Prior art date
Application number
PCT/CN2020/099253
Other languages
French (fr)
Chinese (zh)
Inventor
唐子豪
刘莉红
刘玉宇
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021114623A1 publication Critical patent/WO2021114623A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/53Recognition of crowd images, e.g. recognition of crowd congestion
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/30ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for calculating health indices; for individual health risk assessment
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • This application relates to the field of artificial intelligence image classification, and in particular to a method, device, computer equipment and storage medium for identifying people with deformed spine.
  • the spine is the central axis of the human body. When the spine is severely deformed, it will not only cause abnormal body appearance and motor dysfunction, but also cause cardiopulmonary dysfunction due to thoracic deformities, reduce the quality of life, and seriously affect the physical and mental health of young people. If it is not prevented and detected early, it will not only affect the patient's body shape and appearance, but also may cause abnormal heart and lung function, premature degeneration of the spine, pain, unbalanced trunk, and even death.
  • Spine deformation refers to the horizontal and vertical deformation of the spine. It is usually said that the horizontal deformation becomes scoliosis and the vertical deformation becomes kyphosis.
  • This application provides a method, device, computer equipment, and storage medium for identifying people with spinal deformity, which realizes the automatic identification of the types of people with spinal deformity through the spectral domain method in GCN, and reminds potential people to play a preventive role. Therefore, , Improve the accuracy and reliability of the identification of people with deformed spine, greatly reduce the cost of identification, and serve as a reminder to potential people.
  • a method for identifying people with deformed spine including:
  • the model is a deep convolutional neural network model based on the YOLO model;
  • the enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained
  • input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
  • the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained.
  • Recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
  • a device for identifying people with deformed spine including:
  • the receiving module is used to receive a target recognition instruction, and obtain image data and non-image data associated with the unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is related to the target to be recognized Information;
  • a recognition module configured to input the image data into a back area recognition model, perform back area recognition on the image data through the back area recognition model, and obtain an image of the back area to be recognized that is intercepted by the back area recognition model;
  • the back region recognition model is a deep convolutional neural network model based on the YOLO model;
  • An enhancement module configured to perform image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
  • the acquiring module is configured to input the enhanced image of the region to be recognized into a spine recognition model, extract the features of the spine in the enhanced image of the region to be identified through the spine recognition model, and obtain the output of the spine recognition model according to the features of the spine A first feature vector map, and at the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
  • a filling module configured to perform edge filling of the second feature vector graph to the first feature vector graph to obtain a third feature vector graph
  • An input module configured to input the third feature vector graph into the trained spine graph convolutional network model
  • the output module is used to extract the frequency domain features of the spine in the third feature vector diagram through the spine map convolutional network model according to the spectral domain method in the GCN, and obtain the spine map convolutional network model according to the spine Frequency domain feature output recognition result;
  • the recognition result characterizes the category of the spine deformed population of the target to be recognized, and the category of the spine deformed population includes the scoliosis population, the kyphotic population, the potential scoliosis population, and the potential spine People with hunchbacks and non-spine deformities.
  • a computer device includes a memory, a processor, and a computer program that is stored in the memory and can run on the processor, and the processor implements the following steps when the processor executes the computer program:
  • the back region recognition model is a deep convolutional neural network model based on the YOLO model;
  • the enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained
  • input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
  • the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained.
  • Recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
  • a computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the following steps are implemented:
  • the back region recognition model is a deep convolutional neural network model based on the YOLO model;
  • the enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained
  • input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
  • the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained.
  • Recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
  • This application realizes that according to the captured back image and related non-image information of the target to be recognized, the category of the spine deformed crowd corresponding to the target to be recognized (including potential scoliosis crowds, scoliosis crowds, etc.) can be automatically identified through the spectral domain method in the GCN.
  • People with kyphosis, people with potential scoliosis, people with potential kyphosis, and people without spine deformation can quickly and accurately identify the category of people with spine deformation corresponding to the target to be identified, and remind potential people to play a preventive role Therefore, the accuracy and reliability of the identification of people with deformed spine are improved, the identification cost is greatly reduced, and the potential crowd is reminded.
  • FIG. 1 is a schematic diagram of the application environment of the method for identifying people with deformed spine in an embodiment of the present application
  • FIG. 2 is a flowchart of a method for identifying people with deformed spine in an embodiment of the present application
  • step S20 is a flowchart of step S20 of the method for identifying people with deformed spine in an embodiment of the present application
  • step S30 is a flowchart of step S30 of the method for identifying people with deformed spine in an embodiment of the present application
  • step S40 of the method for identifying people with deformed spine in an embodiment of the present application
  • step S40 of the method for identifying people with deformed spine in another embodiment of the present application.
  • FIG. 7 is a flowchart of step S60 of the method for identifying people with deformed spine in an embodiment of the present application.
  • FIG. 8 is a schematic block diagram of a device for identifying people with deformed spine in an embodiment of the present application.
  • Fig. 9 is a schematic diagram of a computer device in an embodiment of the present application.
  • the technical solution of the present application can be applied to the fields of artificial intelligence, smart city and/or digital medical technology, and can automatically identify people with deformed spine and potential people, and perform disease risk assessment, so as to realize health management and realize smart medical treatment.
  • the method for identifying people with deformed spine can be applied in an application environment as shown in Fig. 1, where a client (computer device) communicates with a server through a network.
  • the client computer equipment
  • the server includes, but is not limited to, various personal computers, notebook computers, smart phones, tablet computers, cameras, and portable wearable devices.
  • the server can be implemented as an independent server or a server cluster composed of multiple servers.
  • a method for identifying people with deformed spine is provided, and the technical solution mainly includes the following steps S10-S70:
  • S10 Receive a target recognition instruction, and obtain image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; and the non-image data is information related to the target to be recognized.
  • the target recognition instruction is an instruction that is triggered when the target to be recognized needs to be recognized
  • the target to be recognized is a person who needs to recognize whether the spine is deformed
  • the unique code is the only one of the target to be recognized.
  • the image data is a photo or image related to the back of the target to be recognized
  • the non-image data is information related to the target to be recognized, for example: the gender and age of the target to be recognized And occupations and other related information.
  • S20 Input the image data into a back area recognition model, and perform back area recognition on the image data through the back area recognition model, and obtain an image of the back area to be recognized that is intercepted by the back area recognition model;
  • the back The region recognition model is a deep convolutional neural network model based on the YOLO model.
  • the back area recognition model is a trained deep convolutional neural network model based on the YOLO model
  • the back area recognition model is a model that recognizes and locates the target to be recognized in the image data.
  • the model of the back area uses the YOLO (You Only Look Once) algorithm to identify the area of the back of the target to be recognized, and the YOLO algorithm uses a CNN (Convolutional Neural Networks, convolutional neural network)
  • the algorithm directly predicts the categories and regions of different targets.
  • the network structure of the YOLO model can be selected according to requirements.
  • the network structure of the YOLO model can be YOLO V1, YOLO V2, YOLO V3, or YOLO V4, etc.
  • the image of the back area to be recognized is an image obtained by intercepting the recognized area after recognition by the back area recognition model. In this way, the back area recognition model can extract only the image of the effective area in the image data to remove interference Image information.
  • the image data is identified by the back region through the back region recognition model, and the back region to be identified extracted by the back region recognition model is obtained.
  • Area image including:
  • the back area recognition model includes the back back area recognition model and the back side area recognition model
  • the image data includes the target back back image and the target back side image.
  • the region recognition model is a trained deep convolutional neural network model based on the YOLO model, and is a model for identifying and locating the back area.
  • the back side area recognition model is the depth of the training based on the YOLO model.
  • the convolutional neural network model is a model that recognizes and locates the back side area.
  • the back image of the target is a photo of the back back when the target to be recognized is wearing underwear or a naked back.
  • the back side image of the target To take a photo of the back side of the target to be identified wearing underwear or naked back.
  • S202 According to the YOLO algorithm, perform recognition through the back back area recognition model, cut out the back back area image to be recognized that contains only the back back area of the target to be recognized, and at the same time perform recognition through the back side area recognition model, and intercept An image of the back side area to be recognized that only contains the back side of the target to be recognized is generated.
  • the back key points in the back image of the target back are identified through the back back area recognition model.
  • the back key points include the left back shoulder point, the right back shoulder point, the upper point of the spine, The midpoint of the spine, the lower point of the spine, the left waist point and the right waist point, locate the position and area of the back according to the identified key points of the back, and intercept the position and area of the located back to obtain the
  • the image of the back area to be identified, the image of the back area to be identified is the back of the back that contains only the object to be identified, and does not include the neck and arms of the object to be identified; according to the YOLO algorithm, through the back side
  • the region recognition model recognizes the side key points in the back side image of the target.
  • the side key points include the upper point of the arm, the middle point of the arm, the lower point of the arm, the neck point, the upper point of the lateral back, the midpoint of the lateral back, and the side
  • the lower point of the back and the front chest point locate the position and area of the side according to the identified key points of the side, and intercept the position and area of the positioned side, so as to obtain the image of the side area of the back to be identified.
  • the image of the back side area to be recognized is the back side only containing the target to be recognized.
  • S203 Determine the image of the back area to be recognized and the image of the back side area to be recognized as the image of the back area to be recognized.
  • the image of the back area to be identified and the image of the side area of the back to be identified are marked as the image of the back area to be identified.
  • This application intercepts the back area image and the side area image of the back of the target to be identified, and can pass the dimensions corresponding to the back and the side of the back, that is, the back corresponds to the horizontal dimension and the side corresponds to the vertical dimension. Intercepting images in multiple dimensions provides an effective image for improving the accuracy and reliability of recognition, and improves the efficiency of recognition.
  • S30 Perform image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified.
  • the image enhancement processing refers to image processing operations such as gray-scale processing, noise removal, and edge enhancement of the back image to be recognized, wherein the gray-scale processing is to process the three colors in the color image ( The brightness of the red, green, and blue) components is used as the gray value of the gray image.
  • the noise removal algorithm can be selected according to requirements.
  • the noise removal algorithm can be selected as spatial domain filtering algorithm, transform domain filtering algorithm, and partial Differential equation algorithm, variational algorithm, morphological noise filtering algorithm, etc., as a preference, the noise removal algorithm is selected as a spatial domain filtering algorithm, and the edge enhancement processing is smoothing the image, and then performing edge points The process of detecting, locating edges, and sharpening the edges.
  • the enhanced image of the area to be recognized is the image obtained after the image enhancement processing.
  • the image enhancement processing can enhance and optimize the to-be-recognized area Features related to spinal deformation in the image of the back area, and can facilitate the recognition of spinal deformation, and improve the accuracy of recognition.
  • step S30 that is, performing image enhancement processing on the image of the back area to be identified to obtain an enhanced image of the area to be identified includes:
  • S301 Perform grayscale processing on the image of the back area to be recognized in the image of the back area to be recognized to obtain a grayscale image of the back back, and at the same time perform a grayscale process on the side of the back to be recognized in the back area image to be recognized.
  • the area image is gray-scaled to obtain a gray-scale image of the back side.
  • the image of the back area to be recognized is separated by channels, and the red channel image of the red channel, the green channel image of the green channel, and the blue channel image of the blue channel are separated, and the back area image to be recognized includes three Image with two channels (red channel, green channel and blue channel), that is, each pixel in the cropped image has three channel component values, which are the red component value, the green component value and the blue component value, Perform grayscale processing on the red channel image, green channel image, and blue channel image to obtain a grayscale image corresponding to the grayscale channel.
  • the red ( The R) component value, the green (G) component value and the blue (B) component value are calculated by the weighted average method to obtain the gray component value of each pixel.
  • S302 Perform image denoising and edge enhancement processing on the back and back grayscale image to obtain a back back enhanced image, and simultaneously perform image denoising and edge enhancement processing on the back side grayscale image to obtain a back side enhanced image.
  • the spatial domain image of the back and back grayscale image is denoised (also known as noise reduction) through a spatial domain filtering algorithm.
  • the spatial domain filtering algorithm can be selected according to requirements, for example, it can be a neighborhood average. Method, median filter, low-pass filter, etc.
  • the spatial domain filter algorithm is selected as the neighborhood average method, that is, the value of each pixel in the back and back grayscale image is adjacent to the periphery of the pixel Take the average of the values of the pixels to obtain the de-noised value corresponding to the pixel to obtain the de-noised back back grayscale image, and perform edge enhancement processing on the de-noised back back grayscale image , So as to obtain the back side enhanced image, the edge enhancement processing is a process of smoothing the image, then detecting edge points, locating the edge, and sharpening the edge. Similarly, the back side The grayscale image undergoes image denoising and edge enhancement processing to obtain the back side enhanced image.
  • S303 Determine the back enhanced image and the back side enhanced image as the to-be-recognized area enhanced image.
  • the back back enhanced image and the back side enhanced image are marked as the to-be-identified area enhanced image.
  • This application performs grayscale processing, denoising and edge enhancement processing on both the back area image and the back side area image to obtain an optimized enhanced image of the area to be identified, which can enhance the deformation of the spine in the image of the back area to be identified.
  • the spine recognition model is a trained deep convolutional neural network that performs the spine feature extraction on the enhanced image of the to-be-recognized region, recognizes according to the spine feature, and outputs the first feature vector image Model
  • the network structure of the spine recognition model can be set according to requirements.
  • the network structure of the spine recognition model can be an Inception V4 network structure, or a VGG16 network structure, etc., and the spine features are related to the spine.
  • the first feature vector map is a series of arrays containing feature vectors corresponding to the spine features output based on the spine features extracted from the enhanced image of the to-be-recognized region Matrix;
  • the data standardization model is a collection model that determines the correlation function between each dimension in the non-image data and the degree of spine deformation and potential risks based on the collected historical non-image data, and the data standardization model can
  • the non-image data is subjected to normalization processing and edge weighting processing, and the normalization processing is to normalize the data of each dimension in the non-image data according to the rules matching each dimension to obtain unified standard data
  • the edge weight processing is to perform weighting processing on the values after the normalization processing according to the edge weight parameters matched with each dimension, that is, the values after the normalization processing are multiplied by the matched edge weight parameters, and the second feature
  • the vector graph is an array matrix obtained after normalization and edge weight processing of the non-image data.
  • the spine features include scoliosis features and kyphosis features; in S40, that is, the spine features in the enhanced image of the region to be recognized are extracted through the spine recognition model , Acquiring the first feature vector diagram output by the spine recognition model according to the features of the spine includes:
  • the spine recognition model includes the scoliosis recognition model and the kyphosis recognition model, and the scoliosis recognition model is trained and trained by extracting the scoliosis features from most of the back images containing the back.
  • the completed neural network model The network structure of the side curve recognition model can be set according to requirements.
  • the hunchback recognition model is a neural network that is trained and trained by extracting the hunchback feature from most of the back side images.
  • Model, the network structure of the humpback recognition model can be set according to requirements, the side bending feature is the asymmetrical two sides of the back trunk, the torso is curved, and the shoulders are uneven.
  • the humpback feature is the convex back and the side back is present.
  • the back back enhanced image is input into the side curve recognition model, the side curve feature in the back back enhanced image is extracted through the side curve recognition model, and the back side enhanced image Input into the hunchback recognition model, and extract the hunchback feature in the back side enhanced image through the hunchback recognition model.
  • S402 Obtain a side curve feature vector diagram output by the side curve recognition model according to the side curve feature, and simultaneously obtain a hunchback feature vector diagram output by the hunchback recognition model according to the hunchback feature.
  • the convolution layer, the pooling layer, and the fully connected layer are processed on the back and back enhanced image through the side curve recognition model to extract the side curve feature, and the extracted side curve feature Arrange and output into a eigenvector diagram, which is the side curve eigenvector diagram.
  • the side curve eigenvector diagram is a matrix diagram containing multiple eigenvectors.
  • the side curve eigenvector diagram is a 100 ⁇ 100 matrix.
  • the graph is the humpback feature vector graph
  • the humpback feature vector graph is a matrix graph containing multiple feature vectors.
  • the humpback feature vector graph is a 100 ⁇ 100 matrix.
  • the side curve eigenvector diagram and the hunchback eigenvector diagram are connected up and down in a matrix to obtain the first eigenvector diagram.
  • the side curve eigenvector diagram is a 100 ⁇ 100 matrix
  • the The humpback feature vector graph is a 100 ⁇ 100 matrix
  • the first feature vector graph is a 100 ⁇ 200 matrix.
  • This application uses the scoliosis recognition model in the spine recognition model to extract the features of the back enhanced image and identify the scoliosis feature vector image, and at the same time, the kyphosis recognition model in the spine recognition model extracts the kyphosis feature of the back side enhanced image
  • the recognition of the humpback feature vector map can more specifically identify the side curvature and the humpback in the spinal deformation, which improves the accuracy and reliability of the recognition.
  • step S40 that is, the normalization and edge weight processing of the non-image data through the data standardization model to obtain the second feature vector graph includes:
  • S404 Acquire each dimension in the non-image data and dimension data corresponding to each of the dimensions.
  • the non-image data is information related to the target to be identified, and the non-image data contains multiple dimensions. These dimensions can be set according to requirements, and the corresponding dimensions of the non-image data are obtained.
  • the dimensional data of the dimensional data is the corresponding input content of the target to be identified for the dimensionality.
  • the dimensions in the non-image data include target gender, target age, target occupation, and target information.
  • the target gender is the gender of the target to be recognized
  • the target age is the age of the target to be recognized
  • the target occupation is the occupation of the target to be recognized
  • the target information is
  • the family information related to the target to be identified may be a family map, and the target information can be set according to requirements.
  • the normalization rule is a rule for unified processing of the dimension data that matches the dimension, that is, the dimension data of the same dimension is converted in a unified rule, for example: the dimension is male in the target age The dimensional data is converted to 1, and the dimensional data of the target age is converted to 0 and other rules.
  • the edge weight parameter is a weighting parameter preset according to the dimension and matched with the dimension, and the edge weight parameter is based on historical data. Obtained from statistical analysis, the edge weight parameter indicates a measurement index of the degree of potential correlation between the dimension and the spine feature.
  • S406 Perform a normalization process on all the dimension data according to the normalization rule matched with each of the dimensions to obtain a dimension standard value corresponding to each of the dimensions.
  • the normalization process is to convert dimensional data into a unified format according to a normalization rule, and perform unified conversion on all the dimensional data according to the normalization rule matching each dimension to obtain The dimension standard value corresponding to each of the dimensions.
  • S407 Perform edge weighting processing on all standard values of the dimensions according to the edge weight parameters matched with each of the dimensions to obtain a weight value corresponding to each of the dimensions.
  • the dimension standard value and the edge weight parameter matching the dimension are multiplied to obtain the weight value, and the edge weighting is processed as multiplying the dimension standard value by the corresponding edge weight
  • the parameter that is, the dimension standard value is expanded according to the edge weight parameter, and the potential correlation degree between each dimension and the spine feature is opened, so that it can be identified objectively and scientifically.
  • the expansion is to copy and fill all the weights to reach a preset matrix size, and the copy and fill is to copy all the weights in one dimension as a whole to the preset matrix size. Finally, the elements that are insufficient to be copied are filled with zeros, and then all the elements of the row are copied to the longitudinal length of the preset matrix size, thereby obtaining the second eigenvector map.
  • This application performs normalization and edge weighting of non-image data and generates a second feature vector diagram related to the features of spine deformation, which can better establish the third feature vector diagram of the relationship between each feature vector, and improve recognition Reliability.
  • S50 Perform edge filling of the second feature vector graph to the first feature vector graph to obtain a third feature vector graph.
  • the edge filling refers to filling on the edge of the array on the basis of the first eigenvector image, that is, filling the second eigenvector image above and below the matrix of the first eigenvector image to the preset value. Set the size to obtain the third feature vector diagram.
  • S60 Input the third feature vector map into the trained spine map convolutional network model.
  • the spine diagram convolutional network model is a neural network model that is identified and trained based on the network structure of the diagram convolution, and the spine diagram convolutional network model can be based on the elements in the third feature vector diagram.
  • the spine frequency domain features are extracted from the feature vector and the association relationship between each element, and the extracted spine frequency domain features are classified and recognized, and finally the category of the corresponding spine deformed crowd is output.
  • the method before step S60, that is, before inputting the third feature vector image into the trained spine graph convolutional network model, the method includes:
  • sample data set includes sample data and sample labels corresponding to the sample data one-to-one; the sample data is a historical third feature vector diagram; the sample labels include scoliosis people and kyphotic people , Potential scoliosis crowd and potential spine kyphosis crowd.
  • the historical third feature vector graph is obtained after the sample image data and sample non-image data associated with the sample data are processed through the above-mentioned steps S20 to S50, and the sample data set contains a plurality of the samples.
  • Data one sample data is associated with one sample label, and the sample label includes a scoliosis crowd, a kyphosis crowd, a potential scoliosis crowd, and a potential kyphosis crowd.
  • S602 Input the sample data into a spinal diagram convolutional neural network model containing initial parameters.
  • the initial parameters of the spine graph convolutional neural network model can be set according to requirements.
  • the initial parameters can be obtained through a transfer learning method to obtain all the parameters of other graph convolution models related to the back recognition, or all Set to a preset value.
  • the GCN is Graph Convolutional Network, that is, to identify the existence or potential classification by extracting relevant spatial features of a topological graph that uses vertices and edges to establish a corresponding relationship.
  • the spectral domain method is by relying on topological
  • the eigenvalues and eigenvectors of the Laplacian matrix corresponding to the graph are used to study the properties of the topological graph and the GCN classification result obtained by Fourier transforming the eigenvalues and eigenvectors of the Laplacian matrix.
  • the frequency domain feature is a feature that is transformed by using the frequency domain method and is related to spinal deformation
  • the spine map convolutional neural network model identifies the sample result of the sample data according to the extracted frequency domain feature of the spine,
  • the sample results include people with scoliosis, people with kyphosis, people with potential scoliosis, people with potential kyphosis, and people without spinal deformity.
  • S604 Determine a loss value according to the sample result and the sample label corresponding to the sample data.
  • the sample result and the sample label are input into the loss function in the spine map convolutional neural network model, and the loss value corresponding to the sample data is calculated.
  • the loss function can be set according to requirements
  • the loss function is the logarithm of the difference between the sample result and the sample label, and indicates the difference between the sample result and the sample label.
  • the convergence condition may be a condition that the loss value is less than the set threshold, that is, when the loss value is less than the set threshold, the spine map convolutional neural network model after convergence is recorded as training completed The convolutional network model of the spine diagram.
  • step S604 that is, after determining the loss value according to the sample result and the sample label corresponding to the sample data, the method further includes:
  • the convergence condition may also be a condition that the loss value is small and will not decrease after 10,000 calculations, that is, the loss value is small and does not fall after 10,000 calculations.
  • the training is stopped, and the spine map convolutional neural network model after convergence is recorded as the spine map convolutional network model that has been trained.
  • the initial parameters of the iterative convolutional neural network model of the spine diagram can be continuously updated, which can continuously move closer to the accurate recognition result, so that the accuracy of the recognition result becomes more and more accurate. high.
  • the spectral domain method in the GCN extract the spine frequency domain features in the third feature vector map through the spine map convolution network model, and obtain the spine map convolution network model according to the spine frequency domain features
  • the output recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes the scoliosis crowd, the kyphotic crowd, the potential scoliosis crowd, the potential kyphotic crowd, and People with non-spine deformities.
  • the GCN is Graph Convolutional Network, that is, to identify the existence or potential classification by extracting relevant spatial features of a topological graph that uses vertices and edges to establish a corresponding relationship.
  • the spectral domain method is by means of topology
  • the eigenvalues and eigenvectors of the Laplacian matrix corresponding to the graph are used to study the properties of the topological graph and the method of performing the Fourier transform on the eigenvalues and eigenvectors of the Laplacian matrix to obtain the result of GCN classification.
  • the spine map convolutional network model recognizes the back photos of the target to be recognized, determines the category of the spine deformed crowd where the target to be recognized is located, can remind potential crowds, and can target the spine deformed after determination The corresponding reminder or prevention reminder is output for the category of the crowd.
  • This application obtains the image data and non-image data associated with the unique code corresponding to the target to be recognized by receiving the target recognition instruction; inputs the image data into the back area recognition model, and obtains the back to be recognized cut out by the back area recognition model Regional image; perform image enhancement processing on the image of the back region to be recognized to obtain an enhanced image of the region to be recognized; input the enhanced image of the region to be recognized into a spine recognition model, and extract the enhanced image of the region to be recognized through the spine recognition model
  • the spine feature in the spine acquiring the first feature vector diagram output by the spine recognition model according to the spine feature, and inputting the non-image data into the data standardization model, and classifying the non-image data through the data standardization model Unification and edge weight processing to obtain the second eigenvector image; edge-filling the second eigenvector image to the first eigenvector image to obtain the third eigenvector image; according to the spectral domain method in GCN, pass
  • the spine map convolution network model extracts the spine frequency
  • This application realizes that by acquiring the image data and non-image data associated with the target to be recognized; intercepting the image of the back area to be recognized in the back area of the image data; performing image enhancement processing on the image of the back area to be recognized to obtain the to be recognized Region-enhanced image; extract the features of the spine in the enhanced image of the region to be identified through the spine recognition model to obtain the first feature vector diagram, and at the same time normalize the non-image data and process the edge weights through the data standardization model to obtain the first Second feature vector graph; edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph; according to the spectral domain method in GCN, the spine graph convolutional network model is used to extract the The frequency domain features of the spine in the third feature vector diagram are obtained, and the recognition results output by the convolutional network model of the spine diagram according to the frequency domain features of the spine are obtained.
  • Non-image information through the spectrum domain method in the GCN, automatically identify the category of the spine deformed population corresponding to the target to be identified (including the potential crowd scoliosis crowd, spine kyphosis crowd, potential scoliosis crowd, potential spine kyphosis crowd and potential scoliosis crowd).
  • People with non-spine deformities can quickly and accurately identify the categories of people with spine deformities corresponding to the target to be identified, and remind potential people to prevent them. Therefore, the accuracy and reliability of the recognition of people with spine deformities are improved. , Which greatly reduces the cost of identification and serves as a reminder to potential crowds.
  • a device for identifying people with deformed spine is provided, and the device for identifying people with deformed spine corresponds one-to-one with the method for identifying people with deformed spine in the above-mentioned embodiment.
  • the device for identifying people with deformed spine includes a receiving module 11, a recognition module 12, an enhancement module 13, an acquisition module 14, a filling module 15, an input module 16 and an output module 17.
  • the detailed description of each functional module is as follows:
  • the receiving module 11 is configured to receive a target recognition instruction, and obtain image data and non-image data associated with the unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is the target to be recognized Related information;
  • the recognition module 12 is configured to input the image data into a back area recognition model, and perform back area recognition on the image data through the back area recognition model, and obtain the back area image to be recognized cut out by the back area recognition model ;
  • the back region recognition model is a deep convolutional neural network model based on the YOLO model;
  • the enhancement module 13 is configured to perform image enhancement processing on the image of the back area to be identified to obtain an enhanced image of the area to be identified;
  • the acquiring module 14 is configured to input the enhanced image of the region to be recognized into a spine recognition model, extract the features of the spine in the enhanced image of the region to be identified through the spine recognition model, and obtain the spine recognition model and output according to the spine features
  • the non-image data is input into a data standardization model, and the non-image data is normalized and edge weighted through the data standardization model to obtain a second feature vector diagram
  • the filling module 15 is used for edge filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
  • the input module 16 is configured to input the third feature vector map into the trained spine map convolutional network model
  • the output module 17 is configured to extract the spine frequency domain features in the third feature vector diagram through the spine map convolutional network model according to the spectral domain method in the GCN, and obtain the spine map convolutional network model according to the spine map convolutional network model.
  • the recognition result of the spine frequency domain feature output; the recognition result characterizes the category of the spine deformed population of the target to be recognized, and the category of the spine deformed population includes the scoliosis population, the kyphotic population, the potential scoliosis population, and the potential scoliosis population. People with kyphosis and non-spine deformities.
  • the various modules in the device for identifying people with deformed spine can be implemented in whole or in part by software, hardware, and a combination thereof.
  • the above-mentioned modules may be embedded in the form of hardware or independent of the processor in the computer equipment, or may be stored in the memory of the computer equipment in the form of software, so that the processor can call and execute the operations corresponding to the above-mentioned modules.
  • a computer device is provided.
  • the computer device may be a server, and its internal structure diagram may be as shown in FIG. 9.
  • the computer equipment includes a processor, a memory, a network interface, and a database connected through a system bus.
  • the processor of the computer device is used to provide calculation and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium stores an operating system, a computer program and a database.
  • the internal memory provides an environment for the operation of the operating system and computer programs in the non-volatile storage medium.
  • the network interface of the computer device is used to communicate with an external terminal through a network connection.
  • the computer program is executed by the processor to realize a method for identifying people with deformed spine.
  • a computer device including a memory, a processor, and a computer program stored in the memory and capable of running on the processor.
  • the processor executes the computer program to implement the method for identifying people with deformed spine in the above-mentioned embodiment. .
  • a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, the method for identifying people with deformed spine in the above-mentioned embodiment is realized.
  • the storage medium involved in this application such as a computer-readable storage medium, may be non-volatile or volatile.
  • Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Biomedical Technology (AREA)
  • Medical Informatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Pathology (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Epidemiology (AREA)
  • Primary Health Care (AREA)
  • Human Computer Interaction (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

Provided are a method, apparatus, computer device, and storage medium for identifying persons having deformed spinal columns, relating to the field of artificial intelligence, said method comprising: obtaining image data and non-image data associated with a target to be identified; extracting an image of a back area to be recognized in a back area in the image data; performing image enhancement processing on the image of the back area to be recognized to obtain an enhanced image of the area to be recognized; extracting the features of the spinal column in the enhanced image of the area to be recognized by means of a spinal column recognition model to obtain a first feature vector map, and performing normalization and edge weight processing of non-image data by means of a data standardization model to obtain a second feature vector map; edge-filling the second feature vector map to the first feature vector map to obtain a third feature vector map; according to the spectral domain method in GCN, extracting the frequency domain features of the spinal column by means of a spinal column image convolutional network model to obtain an identification result. The method can automatically identify the type of persons having a deformed spinal column.

Description

脊柱变形人群识别方法、装置、计算机设备及存储介质Method, device, computer equipment and storage medium for identifying people with deformed spine
本申请要求于2020年6月8日提交中国专利局、申请号为202010513066.6,发明名称为“脊柱变形人群识别方法、装置、计算机设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed with the Chinese Patent Office on June 8, 2020, with the application number 202010513066.6, and the invention title "Methods, devices, computer equipment and storage media for identifying people with spinal deformities", all of which are approved The reference is incorporated in this application.
技术领域Technical field
本申请涉及人工智能的图像分类领域,尤其涉及一种脊柱变形人群识别方法、装置、计算机设备及存储介质。This application relates to the field of artificial intelligence image classification, and in particular to a method, device, computer equipment and storage medium for identifying people with deformed spine.
背景技术Background technique
脊柱是人体的中轴,脊柱变形严重时不仅会造成身体外观异常、运动功能障碍,还可因胸廓畸形而造成心肺功能障碍,降低生活质量,严重影响青少年身心健康的发育。如果不及早预防和发现,不仅影响患者的体型和外观,而且可能造成心肺功能异常,使脊柱过早退变,出现疼痛,躯干不平衡,甚至会导致死亡。脊柱变形指脊柱发生水平和垂直方向的形变,通常称水平方向的形变为侧弯和垂直方向的形变为驼背。The spine is the central axis of the human body. When the spine is severely deformed, it will not only cause abnormal body appearance and motor dysfunction, but also cause cardiopulmonary dysfunction due to thoracic deformities, reduce the quality of life, and seriously affect the physical and mental health of young people. If it is not prevented and detected early, it will not only affect the patient's body shape and appearance, but also may cause abnormal heart and lung function, premature degeneration of the spine, pain, unbalanced trunk, and even death. Spine deformation refers to the horizontal and vertical deformation of the spine. It is usually said that the horizontal deformation becomes scoliosis and the vertical deformation becomes kyphosis.
检查脊柱变形的方法有很多,主要有莫尔图像测量法、X光片测量法、Adams向前弯腰试验等等,发明人意识到,现有技术的方案中都需要人工物理测量检测、而且检测步骤繁琐,导致检测效率低和成本高,特别是X光片测量法会对青少年造成辐射伤害,而且大部分现有方案都是对已经发生脊柱变形之后的患者人群才能检测出来,无法对潜在的人群进行提醒及预防。There are many methods to check spinal deformation, mainly including Moiré image measurement method, X-ray film measurement method, Adams forward bending test, etc. The inventor realized that the existing technical solutions require manual physical measurement and detection, and The detection steps are cumbersome, resulting in low detection efficiency and high cost. In particular, X-ray film measurement can cause radiation damage to teenagers. Most of the existing solutions can only detect the patient population after the spinal deformation has occurred. Remind and prevent the crowd.
发明内容Summary of the invention
本申请提供一种脊柱变形人群识别方法、装置、计算机设备及存储介质,实现了通过GCN中的频谱域方法自动识别出脊柱变形人群的类别,对潜在的人群进行提醒,起到预防作用,因此,提高了脊柱变形人群识别的准确率和可靠性,大大降低了识别成本,并对潜在的人群起到提醒作用。This application provides a method, device, computer equipment, and storage medium for identifying people with spinal deformity, which realizes the automatic identification of the types of people with spinal deformity through the spectral domain method in GCN, and reminds potential people to play a preventive role. Therefore, , Improve the accuracy and reliability of the identification of people with deformed spine, greatly reduce the cost of identification, and serve as a reminder to potential people.
一种脊柱变形人群识别方法,包括:A method for identifying people with deformed spine, including:
接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;Receiving a target recognition instruction, and obtaining image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is information related to the target to be recognized;
将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;Input the image data into a back region recognition model, and perform back region recognition on the image data through the back region recognition model, and obtain the back region image to be recognized cut out by the back region recognition model; the back region recognition The model is a deep convolutional neural network model based on the YOLO model;
对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;Performing image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained At the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;Edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;Input the third feature vector map into the trained spine map convolutional network model;
根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。According to the spectral domain method in GCN, the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained. Recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
一种脊柱变形人群识别装置,包括:A device for identifying people with deformed spine, including:
接收模块,用于接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;The receiving module is used to receive a target recognition instruction, and obtain image data and non-image data associated with the unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is related to the target to be recognized Information;
识别模块,用于将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;A recognition module, configured to input the image data into a back area recognition model, perform back area recognition on the image data through the back area recognition model, and obtain an image of the back area to be recognized that is intercepted by the back area recognition model; The back region recognition model is a deep convolutional neural network model based on the YOLO model;
增强模块,用于对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;An enhancement module, configured to perform image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
获取模块,用于将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The acquiring module is configured to input the enhanced image of the region to be recognized into a spine recognition model, extract the features of the spine in the enhanced image of the region to be identified through the spine recognition model, and obtain the output of the spine recognition model according to the features of the spine A first feature vector map, and at the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
填充模块,用于将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;A filling module, configured to perform edge filling of the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
输入模块,用于将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;An input module, configured to input the third feature vector graph into the trained spine graph convolutional network model;
输出模块,用于根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。The output module is used to extract the frequency domain features of the spine in the third feature vector diagram through the spine map convolutional network model according to the spectral domain method in the GCN, and obtain the spine map convolutional network model according to the spine Frequency domain feature output recognition result; the recognition result characterizes the category of the spine deformed population of the target to be recognized, and the category of the spine deformed population includes the scoliosis population, the kyphotic population, the potential scoliosis population, and the potential spine People with hunchbacks and non-spine deformities.
一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现以下步骤:A computer device includes a memory, a processor, and a computer program that is stored in the memory and can run on the processor, and the processor implements the following steps when the processor executes the computer program:
接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;Receiving a target recognition instruction, and obtaining image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is information related to the target to be recognized;
将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;Input the image data into the back region recognition model, and perform the back region recognition on the image data through the back region recognition model, and obtain the back region image to be recognized cut out by the back region recognition model; the back region recognition The model is a deep convolutional neural network model based on the YOLO model;
对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;Performing image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained At the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;Edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;Input the third feature vector map into the trained spine map convolutional network model;
根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。According to the spectral domain method in GCN, the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained. Recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现以下步骤:A computer-readable storage medium, the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the following steps are implemented:
接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;Receiving a target recognition instruction, and obtaining image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is information related to the target to be recognized;
将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;Input the image data into the back region recognition model, and perform the back region recognition on the image data through the back region recognition model, and obtain the back region image to be recognized cut out by the back region recognition model; the back region recognition The model is a deep convolutional neural network model based on the YOLO model;
对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;Performing image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识 别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained At the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;Edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;Input the third feature vector map into the trained spine map convolutional network model;
根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。According to the spectral domain method in GCN, the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained. Recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
本申请实现了根据待识别目标的拍摄的背部的图像和相关的非图像信息,通过GCN中的频谱域方法自动识别出待识别目标对应的脊柱变形人群的类别(包括潜在的人群侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群),能够快速地、准确地识别出待识别目标对应脊柱变形人群的类别,对潜在的人群进行提醒,起到预防作用,因此,提高了脊柱变形人群识别的准确率和可靠性,大大降低了识别成本,并对潜在的人群起到提醒作用。This application realizes that according to the captured back image and related non-image information of the target to be recognized, the category of the spine deformed crowd corresponding to the target to be recognized (including potential scoliosis crowds, scoliosis crowds, etc.) can be automatically identified through the spectral domain method in the GCN. People with kyphosis, people with potential scoliosis, people with potential kyphosis, and people without spine deformation) can quickly and accurately identify the category of people with spine deformation corresponding to the target to be identified, and remind potential people to play a preventive role Therefore, the accuracy and reliability of the identification of people with deformed spine are improved, the identification cost is greatly reduced, and the potential crowd is reminded.
附图说明Description of the drawings
图1是本申请一实施例中脊柱变形人群识别方法的应用环境示意图;FIG. 1 is a schematic diagram of the application environment of the method for identifying people with deformed spine in an embodiment of the present application;
图2是本申请一实施例中脊柱变形人群识别方法的流程图;2 is a flowchart of a method for identifying people with deformed spine in an embodiment of the present application;
图3是本申请一实施例中脊柱变形人群识别方法的步骤S20的流程图;3 is a flowchart of step S20 of the method for identifying people with deformed spine in an embodiment of the present application;
图4是本申请一实施例中脊柱变形人群识别方法的步骤S30的流程图;4 is a flowchart of step S30 of the method for identifying people with deformed spine in an embodiment of the present application;
图5是本申请一实施例中脊柱变形人群识别方法的步骤S40的流程图;5 is a flowchart of step S40 of the method for identifying people with deformed spine in an embodiment of the present application;
图6是本申请另一实施例中脊柱变形人群识别方法的步骤S40的流程图;6 is a flowchart of step S40 of the method for identifying people with deformed spine in another embodiment of the present application;
图7是本申请一实施例中脊柱变形人群识别方法的步骤S60的流程图;FIG. 7 is a flowchart of step S60 of the method for identifying people with deformed spine in an embodiment of the present application;
图8是本申请一实施例中脊柱变形人群识别装置的原理框图;8 is a schematic block diagram of a device for identifying people with deformed spine in an embodiment of the present application;
图9是本申请一实施例中计算机设备的示意图。Fig. 9 is a schematic diagram of a computer device in an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行描述。The technical solutions in the embodiments of the present application will be described below in conjunction with the drawings in the embodiments of the present application.
本申请的技术方案可应用于人工智能、智慧城市和/或数字医疗技术领域,可自动识别出脊柱变形人群及潜在人群,进行疾病风险评估,以实现健康管理,实现智慧医疗。The technical solution of the present application can be applied to the fields of artificial intelligence, smart city and/or digital medical technology, and can automatically identify people with deformed spine and potential people, and perform disease risk assessment, so as to realize health management and realize smart medical treatment.
本申请提供的脊柱变形人群识别方法,可应用在如图1的应用环境中,其中,客户端(计算机设备)通过网络与服务器进行通信。其中,客户端(计算机设备)包括但不限于为各种个人计算机、笔记本电脑、智能手机、平板电脑、摄像头和便携式可穿戴设备。服务器可以用独立的服务器或者是多个服务器组成的服务器集群来实现。The method for identifying people with deformed spine provided in this application can be applied in an application environment as shown in Fig. 1, where a client (computer device) communicates with a server through a network. Among them, the client (computer equipment) includes, but is not limited to, various personal computers, notebook computers, smart phones, tablet computers, cameras, and portable wearable devices. The server can be implemented as an independent server or a server cluster composed of multiple servers.
在一实施例中,如图2所示,提供一种脊柱变形人群识别方法,其技术方案主要包括以下步骤S10-S70:In one embodiment, as shown in FIG. 2, a method for identifying people with deformed spine is provided, and the technical solution mainly includes the following steps S10-S70:
S10,接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息。S10: Receive a target recognition instruction, and obtain image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; and the non-image data is information related to the target to be recognized.
可理解地,所述目标识别指令为需要对所述待识别目标进行识别时触发的指令,所述待识别目标为需要识别脊柱是否变形的人,所述唯一码为所述待识别目标的唯一的标识码,所述图像数据为拍摄所述待识别目标的背部相关的照片或者图像,所述非图像数据为与所述待识别目标相关的信息,例如:所述待识别目标的性别、年龄和职业等相关信息。Understandably, the target recognition instruction is an instruction that is triggered when the target to be recognized needs to be recognized, the target to be recognized is a person who needs to recognize whether the spine is deformed, and the unique code is the only one of the target to be recognized. The image data is a photo or image related to the back of the target to be recognized, and the non-image data is information related to the target to be recognized, for example: the gender and age of the target to be recognized And occupations and other related information.
S20,将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型。S20: Input the image data into a back area recognition model, and perform back area recognition on the image data through the back area recognition model, and obtain an image of the back area to be recognized that is intercepted by the back area recognition model; the back The region recognition model is a deep convolutional neural network model based on the YOLO model.
可理解地,所述背部区域识别模型为训练完成的基于YOLO模型搭架的深度卷积神经网络模型,所述背部区域识别模型为识别并定位出所述图像数据中的所述待识别目标的背部区域的模型,所述背部区域识别模型通过YOLO(You Only Look Once)算法识别出所述待识别目标的背部的区域,所述YOLO算法为使用一个CNN(Convolutional Neural Networks,卷积神经网络)运算直接预测不同目标的类别和区域的算法,所述YOLO模型的网络结构可以根据需求进行选取,比如所述YOLO模型的网络结构可以为YOLO V1、YOLO V2、YOLO V3或者YOLO V4等,所述待识别背部区域图像为通过所述背部区域识别模型识别后将识别后的区域进行截取获得的图像,如此,通过所述背部区域识别模型能够只提取所述图像数据中有效区域的图像,去除干扰的图像信息。Understandably, the back area recognition model is a trained deep convolutional neural network model based on the YOLO model, and the back area recognition model is a model that recognizes and locates the target to be recognized in the image data. The model of the back area, the back area recognition model uses the YOLO (You Only Look Once) algorithm to identify the area of the back of the target to be recognized, and the YOLO algorithm uses a CNN (Convolutional Neural Networks, convolutional neural network) The algorithm directly predicts the categories and regions of different targets. The network structure of the YOLO model can be selected according to requirements. For example, the network structure of the YOLO model can be YOLO V1, YOLO V2, YOLO V3, or YOLO V4, etc. The image of the back area to be recognized is an image obtained by intercepting the recognized area after recognition by the back area recognition model. In this way, the back area recognition model can extract only the image of the effective area in the image data to remove interference Image information.
在一实施例中,如图3所示,所述步骤S20中,即通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像,包括:In an embodiment, as shown in FIG. 3, in the step S20, the image data is identified by the back region through the back region recognition model, and the back region to be identified extracted by the back region recognition model is obtained. Area image, including:
S201,将目标背部背面图像输入所述背部区域识别模型中的背部背面区域识别模型,同时将目标背部侧面图像输入所述背部区域识别模型中的背部侧面区域识别模型;所述图像数据包括所述目标背部背面图像和所述目标背部侧面图像。S201. Input the back image of the target back into the back back area recognition model in the back area recognition model, and input the back side image of the target into the back side area recognition model in the back area recognition model; the image data includes the An image of the back of the target and a side image of the back of the target.
可理解地,所述背部区域识别模型包括所述背部背面区域识别模型和所述背部侧面区域识别模型,所述图像数据包括所述目标背部背面图像和所述目标背部侧面图像,所述背部背面区域识别模型为训练完成的基于YOLO模型搭架的深度卷积神经网络模型,且为识别并定位出背部背面区域的模型,所述背部侧面区域识别模型为训练完成的基于YOLO模型搭架的深度卷积神经网络模型,且为识别并定位出背部侧面区域的模型,所述目标背部背面图像为拍摄所述待识别目标穿着内衣或者裸背情况下的背部背面的照片,所述目标背部侧面图像为拍摄所述待识别目标穿着内衣或者裸背情况下的背部侧面的照片。Understandably, the back area recognition model includes the back back area recognition model and the back side area recognition model, and the image data includes the target back back image and the target back side image. The region recognition model is a trained deep convolutional neural network model based on the YOLO model, and is a model for identifying and locating the back area. The back side area recognition model is the depth of the training based on the YOLO model. The convolutional neural network model is a model that recognizes and locates the back side area. The back image of the target is a photo of the back back when the target to be recognized is wearing underwear or a naked back. The back side image of the target To take a photo of the back side of the target to be identified wearing underwear or naked back.
S202,根据YOLO算法,通过所述背部背面区域识别模型进行识别,截取出只含有所述待识别目标的背部背面的待识别背部背面区域图像,同时通过所述背部侧面区域识别模型进行识别,截取出只含有所述待识别目标的背部侧面的待识别背部侧面区域图像。S202. According to the YOLO algorithm, perform recognition through the back back area recognition model, cut out the back back area image to be recognized that contains only the back back area of the target to be recognized, and at the same time perform recognition through the back side area recognition model, and intercept An image of the back side area to be recognized that only contains the back side of the target to be recognized is generated.
可理解地,根据YOLO算法,通过所述背部背面区域识别模型识别出所述目标背部背面图像中背面关键点,所述背面关键点包括左背肩点、右背肩点、背脊的上点、背脊的中点、背脊的下点、左腰点和右腰点,根据识别出的所述背面关键点定位出背面的位置及区域,对定位出背面的位置及区域进行截取,从而获取到所述待识别背部背面区域图像,所述待识别背部背面区域图像为仅含有所述待识别目标的背部背面,不包含有所述待识别目标的脖子和手臂;根据YOLO算法,通过所述背部侧面区域识别模型识别出所述目标背部侧面图像中侧面关键点,所述侧面关键点包括手臂上点、手臂中点、手臂下点、脖子点、侧背脊的上点、侧背脊的中点、侧背脊的下点和前胸点,根据识别出的所述侧面关键点定位出侧面的位置及区域,对定位出侧面的位置及区域进行截取,从而获取到所述待识别背部侧面区域图像,所述待识别背部侧面区域图像为仅含有所述待识别目标的背部侧面。Understandably, according to the YOLO algorithm, the back key points in the back image of the target back are identified through the back back area recognition model. The back key points include the left back shoulder point, the right back shoulder point, the upper point of the spine, The midpoint of the spine, the lower point of the spine, the left waist point and the right waist point, locate the position and area of the back according to the identified key points of the back, and intercept the position and area of the located back to obtain the The image of the back area to be identified, the image of the back area to be identified is the back of the back that contains only the object to be identified, and does not include the neck and arms of the object to be identified; according to the YOLO algorithm, through the back side The region recognition model recognizes the side key points in the back side image of the target. The side key points include the upper point of the arm, the middle point of the arm, the lower point of the arm, the neck point, the upper point of the lateral back, the midpoint of the lateral back, and the side The lower point of the back and the front chest point, locate the position and area of the side according to the identified key points of the side, and intercept the position and area of the positioned side, so as to obtain the image of the side area of the back to be identified. The image of the back side area to be recognized is the back side only containing the target to be recognized.
S203,将所述待识别背部背面区域图像和所述待识别背部侧面区域图像确定为所述待识别背部区域图像。S203: Determine the image of the back area to be recognized and the image of the back side area to be recognized as the image of the back area to be recognized.
可理解地,将所述待识别背部背面区域图像和所述待识别背部侧面区域图像标记为所述待识别背部区域图像。Understandably, the image of the back area to be identified and the image of the side area of the back to be identified are marked as the image of the back area to be identified.
本申请通过截取出待识别目标的背部的背面区域图像和侧面区域图像,能够通过背部的背面和侧面对应的维度,即背面对应为水平维度和侧面对应为垂直维度,对待识别目标的背部的两个维度进行截取图像,为提高识别的准确率和可靠性提供了有效的图像,提高了识别的效率。This application intercepts the back area image and the side area image of the back of the target to be identified, and can pass the dimensions corresponding to the back and the side of the back, that is, the back corresponds to the horizontal dimension and the side corresponds to the vertical dimension. Intercepting images in multiple dimensions provides an effective image for improving the accuracy and reliability of recognition, and improves the efficiency of recognition.
S30,对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像。S30: Perform image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified.
可理解地,所述图像增强处理指对所述待识别背部图像进灰度化处理、去噪音和边缘增强等图像处理操作,其中,所述灰度化处理为将彩色图像中的三色(红、绿、蓝)分量的亮度作为灰度图像的灰度值,所述去噪音的算法可以根据需求进行选取,比如去噪音处理的算法可以选取为空间域滤波算法、变换域滤波算法、偏微分方程算法、变分算法和形态学噪声滤除算法等等,作为优选,所述去噪音的算法选取为空间域滤波算法,所述边缘增强处理为对图像进行平滑处理,再对边缘点进行检测,定位出边缘,将边缘进行锐化的处理过程,所述待识别区域增强图像为经过所述图像增强处理之后获得的图像,如此,通过所述图像增强处理可以增强和优化所述待识别背部区域图像中与脊柱变形相关的特征,并且能够便于脊柱变形的识别,提高识别的准确率。Understandably, the image enhancement processing refers to image processing operations such as gray-scale processing, noise removal, and edge enhancement of the back image to be recognized, wherein the gray-scale processing is to process the three colors in the color image ( The brightness of the red, green, and blue) components is used as the gray value of the gray image. The noise removal algorithm can be selected according to requirements. For example, the noise removal algorithm can be selected as spatial domain filtering algorithm, transform domain filtering algorithm, and partial Differential equation algorithm, variational algorithm, morphological noise filtering algorithm, etc., as a preference, the noise removal algorithm is selected as a spatial domain filtering algorithm, and the edge enhancement processing is smoothing the image, and then performing edge points The process of detecting, locating edges, and sharpening the edges. The enhanced image of the area to be recognized is the image obtained after the image enhancement processing. In this way, the image enhancement processing can enhance and optimize the to-be-recognized area Features related to spinal deformation in the image of the back area, and can facilitate the recognition of spinal deformation, and improve the accuracy of recognition.
在一实施例中,如图4所示,所述步骤S30中,即所述对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像,包括:In one embodiment, as shown in FIG. 4, in the step S30, that is, performing image enhancement processing on the image of the back area to be identified to obtain an enhanced image of the area to be identified includes:
S301,对所述待识别背部区域图像中的所述待识别背部背面区域图像进行灰度化处理,得到背部背面灰度图像,同时对所述待识别背部区域图像中的所述待识别背部侧面区域图像进行灰度化处理,得到背部侧面灰度图像。S301: Perform grayscale processing on the image of the back area to be recognized in the image of the back area to be recognized to obtain a grayscale image of the back back, and at the same time perform a grayscale process on the side of the back to be recognized in the back area image to be recognized. The area image is gray-scaled to obtain a gray-scale image of the back side.
可理解地,将所述待识别背部区域图像通过通道分离,分离出红色通道的红色通道图像、绿色通道的绿色通道图像和蓝色通道的蓝色通道图像,所述待识别背部区域图像包括三个通道(红色通道、绿色通道和蓝色通道)图像,即每个所述裁切图像中的每个像素点有三个通道分量值,分别为红色分量值、绿色分量值和蓝色分量值,对所述红色通道图像、绿色通道图像和蓝色通道图像进行灰度化处理,得到与灰度通道对应的灰度图像,将所述待识别背部区域图像中的每个像素点对应的红色(R)分量值、绿色(G)分量值和蓝色(B)分量值通过加权平均法计算得出每个像素点的灰度分量值,所述加权平均法中的公式可以根据需求设定,比如加权平均法的公式设定为:Y=0.299R+0.587G+0.114B,其中,Y为每个像素点的灰度分量值,R为每个像素点中的红色分量值,G为每个像素点中的绿色分量值,B为每个像素点中的蓝色分量值,从而得到所述待识别背部区域图像的所述背部背面灰度图像,同理,对所述待识别背部区域图像中的所述待识别背部侧面区域图像进行灰度化处理,从而得到所述背部侧面灰度图像。Understandably, the image of the back area to be recognized is separated by channels, and the red channel image of the red channel, the green channel image of the green channel, and the blue channel image of the blue channel are separated, and the back area image to be recognized includes three Image with two channels (red channel, green channel and blue channel), that is, each pixel in the cropped image has three channel component values, which are the red component value, the green component value and the blue component value, Perform grayscale processing on the red channel image, green channel image, and blue channel image to obtain a grayscale image corresponding to the grayscale channel. The red ( The R) component value, the green (G) component value and the blue (B) component value are calculated by the weighted average method to obtain the gray component value of each pixel. The formula in the weighted average method can be set according to requirements, For example, the formula of the weighted average method is set as: Y=0.299R+0.587G+0.114B, where Y is the gray component value of each pixel, R is the red component value of each pixel, and G is each pixel. The green component value in each pixel, B is the blue component value in each pixel, so as to obtain the back back grayscale image of the back area image to be recognized. Similarly, the back area to be recognized The image of the back side area to be identified in the image is subjected to grayscale processing, so as to obtain the back side grayscale image.
S302,对所述背部背面灰度图像进行图像去噪音及边缘增强处理,得到背部背面增强图像,同时对所述背部侧面灰度图像进行图像去噪音及边缘增强处理,得到背部侧面增强图像。S302: Perform image denoising and edge enhancement processing on the back and back grayscale image to obtain a back back enhanced image, and simultaneously perform image denoising and edge enhancement processing on the back side grayscale image to obtain a back side enhanced image.
可理解地,通过空间域滤波算法,对所述背部背面灰度图像空间域图像进行去噪音(也称为降噪),所述空间域滤波算法可以根据需求选定,比如可以为邻域平均法、中值滤波、低通滤波等,作为优选,所述空间域滤波算法选定为邻域平均法,即将所述背部背面灰度图像中的每个像素点的值与该像素点四周邻近的像素点的值取平均,得到该像素点对应的去噪音后的值,从而得到去噪音后的所述背部背面灰度图像,将去噪音后的所述背部背面灰度图像进行边缘增强处理,从而得到所述背部背面增强图像,所述边缘增强处理为对图像进行平滑处理,再对边缘点进行检测,定位出边缘,将边缘进行锐化的处理过程,同理,对所述背部侧面灰度图像进行图像去噪音及边缘增强处理,得到所述背部侧面增强图像。Understandably, the spatial domain image of the back and back grayscale image is denoised (also known as noise reduction) through a spatial domain filtering algorithm. The spatial domain filtering algorithm can be selected according to requirements, for example, it can be a neighborhood average. Method, median filter, low-pass filter, etc. Preferably, the spatial domain filter algorithm is selected as the neighborhood average method, that is, the value of each pixel in the back and back grayscale image is adjacent to the periphery of the pixel Take the average of the values of the pixels to obtain the de-noised value corresponding to the pixel to obtain the de-noised back back grayscale image, and perform edge enhancement processing on the de-noised back back grayscale image , So as to obtain the back side enhanced image, the edge enhancement processing is a process of smoothing the image, then detecting edge points, locating the edge, and sharpening the edge. Similarly, the back side The grayscale image undergoes image denoising and edge enhancement processing to obtain the back side enhanced image.
S303,将所述背部背面增强图像和所述背部侧面增强图像确定为所述待识别区域增强图像。S303: Determine the back enhanced image and the back side enhanced image as the to-be-recognized area enhanced image.
可理解地,将所述背部背面增强图像和所述背部侧面增强图像标记为所述待识别区域增强图像。Understandably, the back back enhanced image and the back side enhanced image are marked as the to-be-identified area enhanced image.
本申请通过对背部背面区域图像和背部侧面区域图像都进行灰度化处理及去噪和边缘增强处理,能够获得优化后的待识别区域增强图像,起到增强待识别背部区域图像中与脊柱变形相关的特征,并且能够便于脊柱变形的识别,提高识别的准确率。This application performs grayscale processing, denoising and edge enhancement processing on both the back area image and the back side area image to obtain an optimized enhanced image of the area to be identified, which can enhance the deformation of the spine in the image of the back area to be identified. Related features, and can facilitate the recognition of spinal deformation, improve the accuracy of recognition.
S40,将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图。S40. Input the enhanced image of the to-be-recognized area into a spine recognition model, extract the spine features in the enhanced image of the to-be-recognized area through the spine recognition model, and obtain a first feature output by the spine recognition model according to the spine feature A vector map is used to input the non-image data into a data standardization model, and the non-image data is normalized and edge weighted through the data standardization model to obtain a second feature vector map.
可理解地,所述脊柱识别模型为训练完成的对所述待识别区域增强图像进行所述脊柱特征提取并根据所述脊柱特征进行识别及输出所述第一特征向量图的深度卷积神经网络模型,所述脊柱识别模型的网络结构可以根据需求进行设定,比如所述脊柱识别模型的网络结构可以为Inception V4网络结构,也可以为VGG16网络结构等等,所述脊柱特征为与脊柱的形状及弯曲等相关的特征向量,所述第一特征向量图为根据从所述待识别区域增强图像中提取出的所述脊柱特征输出的一系列含有与所述脊柱特征对应的特征向量的数组矩阵;所述数据标准化模型为根据收集的历史的非图像数据而确定出所述非图像数据中的各个维度与脊柱变形程度及潜在风险存在的相关函数的集合模型,所述数据标准化模型能够对所述非图像数据进行归一化处理和边缘权重处理,所述归一化处理为将所述非图像数据中的各个维度的数据根据与各个维度匹配的规则进行数值标准化,得到统一标准的数据,所述边缘权重处理为根据与各个维度匹配的边缘权重参数对归一化处理之后的数值进行加权处理,即将归一化处理之后的数值与匹配的边缘权重参数相乘,所述第二特征向量图为对所述非图像数据进行归一化及边缘权重处理之后获得的数组矩阵。Understandably, the spine recognition model is a trained deep convolutional neural network that performs the spine feature extraction on the enhanced image of the to-be-recognized region, recognizes according to the spine feature, and outputs the first feature vector image Model, the network structure of the spine recognition model can be set according to requirements. For example, the network structure of the spine recognition model can be an Inception V4 network structure, or a VGG16 network structure, etc., and the spine features are related to the spine. Feature vectors related to shape, curvature, etc., the first feature vector map is a series of arrays containing feature vectors corresponding to the spine features output based on the spine features extracted from the enhanced image of the to-be-recognized region Matrix; the data standardization model is a collection model that determines the correlation function between each dimension in the non-image data and the degree of spine deformation and potential risks based on the collected historical non-image data, and the data standardization model can The non-image data is subjected to normalization processing and edge weighting processing, and the normalization processing is to normalize the data of each dimension in the non-image data according to the rules matching each dimension to obtain unified standard data The edge weight processing is to perform weighting processing on the values after the normalization processing according to the edge weight parameters matched with each dimension, that is, the values after the normalization processing are multiplied by the matched edge weight parameters, and the second feature The vector graph is an array matrix obtained after normalization and edge weight processing of the non-image data.
在一实施例中,如图5所示,所述脊柱特征包括侧弯特征和驼背特征;所述S40中,即所述通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,包括:In one embodiment, as shown in FIG. 5, the spine features include scoliosis features and kyphosis features; in S40, that is, the spine features in the enhanced image of the region to be recognized are extracted through the spine recognition model , Acquiring the first feature vector diagram output by the spine recognition model according to the features of the spine includes:
S401,通过侧弯识别模型对所述背部背面增强图像进行所述侧弯特征提取,同时通过驼背识别模型对所述背部侧面增强图像进行所述驼背特征提取;所述脊柱识别模型包括所述侧弯识别模型和所述驼背识别模型。S401. Perform the side curve feature extraction on the back enhanced image using a side curve recognition model, and simultaneously perform the kyphosis feature extraction on the back side enhanced image using a kyphosis recognition model; the spine recognition model includes the side A bend recognition model and the humpback recognition model.
可理解地,所述脊柱识别模型包括所述侧弯识别模型和所述驼背识别模型,所述侧弯识别模型为通过从大多数含有背部背面图像中提取出所述侧弯特征进行训练并且训练完成的神经网络模型,所述侧弯识别模型的网络结构可以根据需求设定,所述驼背识别模型为通过从大多数含有背部侧面图像中提取出所述驼背特征进行训练并且训练完成的神经网络模型,所述驼背识别模型的网络结构可以根据需求设定,所述侧弯特征为背部躯干两侧不对称,躯干弯曲,双肩不平的特征,所述驼背特征为后背凸起,侧背呈现弓形的特征,将所述背部背面增强图像输入所述侧弯识别模型中,通过所述侧弯识别模型提取所述背部背面增强图像中的所述侧弯特征,同时将所述背部侧面增强图像输入所述驼背识别模型中,通过所述驼背识别模型提取所述背部侧面增强图像中的所述驼背特征。Understandably, the spine recognition model includes the scoliosis recognition model and the kyphosis recognition model, and the scoliosis recognition model is trained and trained by extracting the scoliosis features from most of the back images containing the back. The completed neural network model. The network structure of the side curve recognition model can be set according to requirements. The hunchback recognition model is a neural network that is trained and trained by extracting the hunchback feature from most of the back side images. Model, the network structure of the humpback recognition model can be set according to requirements, the side bending feature is the asymmetrical two sides of the back trunk, the torso is curved, and the shoulders are uneven. The humpback feature is the convex back and the side back is present. Arch-shaped feature, the back back enhanced image is input into the side curve recognition model, the side curve feature in the back back enhanced image is extracted through the side curve recognition model, and the back side enhanced image Input into the hunchback recognition model, and extract the hunchback feature in the back side enhanced image through the hunchback recognition model.
S402,获取所述侧弯识别模型根据所述侧弯特征输出的侧弯特征向量图,同时获取所述驼背识别模型根据所述驼背特征输出的驼背特征向量图。S402: Obtain a side curve feature vector diagram output by the side curve recognition model according to the side curve feature, and simultaneously obtain a hunchback feature vector diagram output by the hunchback recognition model according to the hunchback feature.
可理解地,通过所述侧弯识别模型对所述背部背面增强图像进行卷积层、池化层和全连接层的处理后提取出所述侧弯特征,将提取后的所述侧弯特征进行排列输出成特征向量图,即为所述侧弯特征向量图,所述侧弯特征向量图为一个含有多个特征向量的矩阵图,例如所述侧弯特征向量图为100×100的矩阵,同时通过所述驼背识别模型对所述背部侧面增强图像进行卷积层、池化层和全连接层的处理后提取出所述驼背特征,将提取后的所述驼背特征进行输出成特征向量图,即为所述驼背特征向量图,所述驼背特征向量图为一个含有多个特征向量的矩阵图,例如所述驼背特征向量图为100×100的矩阵。Understandably, the convolution layer, the pooling layer, and the fully connected layer are processed on the back and back enhanced image through the side curve recognition model to extract the side curve feature, and the extracted side curve feature Arrange and output into a eigenvector diagram, which is the side curve eigenvector diagram. The side curve eigenvector diagram is a matrix diagram containing multiple eigenvectors. For example, the side curve eigenvector diagram is a 100×100 matrix. , At the same time, perform convolutional layer, pooling layer, and fully connected layer processing on the back side enhanced image through the humpback recognition model to extract the humpback feature, and then output the extracted humpback feature into a feature vector The graph is the humpback feature vector graph, and the humpback feature vector graph is a matrix graph containing multiple feature vectors. For example, the humpback feature vector graph is a 100×100 matrix.
S403,将所述侧弯特征向量图与所述驼背特征向量图进行拼接,得到所述第一特征向量图。S403, splicing the side curve feature vector graph and the humpback feature vector graph to obtain the first feature vector graph.
可理解地,将所述侧弯特征向量图与所述驼背特征向量图进行上下矩阵连接,得到所 述第一特征向量图,例如所述侧弯特征向量图为100×100的矩阵,所述驼背特征向量图为100×100的矩阵,则所述第一特征向量图为100×200的矩阵。Understandably, the side curve eigenvector diagram and the hunchback eigenvector diagram are connected up and down in a matrix to obtain the first eigenvector diagram. For example, the side curve eigenvector diagram is a 100×100 matrix, and the The humpback feature vector graph is a 100×100 matrix, and the first feature vector graph is a 100×200 matrix.
本申请通过脊柱识别模型中的侧弯识别模型对背部背面增强图像进行测完特征提取及识别出侧弯特征向量图,同时通过脊柱识别模型中的驼背识别模型对背部侧面增强图像进行驼背特征提取及识别出驼背特征向量图,能够更具针对性地识别出脊柱变形中的侧弯和驼背,提高了识别的准确率及可靠性。This application uses the scoliosis recognition model in the spine recognition model to extract the features of the back enhanced image and identify the scoliosis feature vector image, and at the same time, the kyphosis recognition model in the spine recognition model extracts the kyphosis feature of the back side enhanced image And the recognition of the humpback feature vector map can more specifically identify the side curvature and the humpback in the spinal deformation, which improves the accuracy and reliability of the recognition.
在一实施例中,如图6所示,所述步骤S40中,即所述通过数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图,包括:In an embodiment, as shown in FIG. 6, in the step S40, that is, the normalization and edge weight processing of the non-image data through the data standardization model to obtain the second feature vector graph includes:
S404,获取所述非图像数据中的各个维度和与各个所述维度对应的维度数据。S404: Acquire each dimension in the non-image data and dimension data corresponding to each of the dimensions.
可理解地,所述非图像数据为与所述待识别目标相关的信息,所述非图像数据中包含多个维度,这些维度可以根据需求设定,获取所述非图像数据中的各个维度对应的维度数据,所述维度数据为所述待识别目标针对该维度对应输入的内容。Understandably, the non-image data is information related to the target to be identified, and the non-image data contains multiple dimensions. These dimensions can be set according to requirements, and the corresponding dimensions of the non-image data are obtained. The dimensional data of the dimensional data is the corresponding input content of the target to be identified for the dimensionality.
在一实施例中,所述非图像数据中的维度包括目标性别、目标年龄、目标职业和目标信息。In an embodiment, the dimensions in the non-image data include target gender, target age, target occupation, and target information.
可理解地,所述目标性别为所述待识别目标的性别,所述目标年龄为所述待识别目标的年龄,所述目标职业为所述待识别目标从事的职业,所述目标信息为所述待识别目标相关的家族信息或者为家族图谱,所述目标信息可以根据需求设定。Understandably, the target gender is the gender of the target to be recognized, the target age is the age of the target to be recognized, the target occupation is the occupation of the target to be recognized, and the target information is The family information related to the target to be identified may be a family map, and the target information can be set according to requirements.
S405,获取与各个所述维度匹配的归一化规则和边缘权重参数。S405: Obtain normalization rules and edge weight parameters that match each of the dimensions.
可理解地,所述归一化规则为将与维度匹配的所述维度数据进行统一化处理的规则,即将相同维度的维度数据进行统一规则的转换,例如:将维度为目标年龄中为男的维度数据转换为1,将维度为目标年龄中为女的维度数据转换为0等规则,所述边缘权重参数为根据维度而预设的与维度匹配的加权参数,所述边缘权重参数根据历史的统计中分析获得,所述边缘权重参数表明了该维度与所述脊柱特征的潜在关联程度的衡量指标。Understandably, the normalization rule is a rule for unified processing of the dimension data that matches the dimension, that is, the dimension data of the same dimension is converted in a unified rule, for example: the dimension is male in the target age The dimensional data is converted to 1, and the dimensional data of the target age is converted to 0 and other rules. The edge weight parameter is a weighting parameter preset according to the dimension and matched with the dimension, and the edge weight parameter is based on historical data. Obtained from statistical analysis, the edge weight parameter indicates a measurement index of the degree of potential correlation between the dimension and the spine feature.
S406,根据与各个所述维度匹配的归一化规则,对所有所述维度数据进行归一化处理,得到与各个所述维度对应的维度标准值。S406: Perform a normalization process on all the dimension data according to the normalization rule matched with each of the dimensions to obtain a dimension standard value corresponding to each of the dimensions.
可理解地,所述归一化处理为根据归一化规则将维度数据转换成统一格式,根据与各所述维度匹配的所述归一化规则,将所有所述维度数据进行统一转换,得到与各个所述维度对应的维度标准值。Understandably, the normalization process is to convert dimensional data into a unified format according to a normalization rule, and perform unified conversion on all the dimensional data according to the normalization rule matching each dimension to obtain The dimension standard value corresponding to each of the dimensions.
S407,根据与各个所述维度匹配的边缘权重参数,对所有与所述维度标准值进行边缘加权处理,得到与各个所述维度对应的加权值。S407: Perform edge weighting processing on all standard values of the dimensions according to the edge weight parameters matched with each of the dimensions to obtain a weight value corresponding to each of the dimensions.
可理解地,将所述维度标准值和与该维度匹配的所述边缘权重参数相乘,得到所述加权值,所述边缘加权处理为将所述维度标准值乘于对应的所述边缘权重参数,即根据所述边缘权重参数对所述维度标准值进行扩大处理,拉开各维度之间与所述脊柱特征潜在关联程度,能够客观地、科学地进行识别。Understandably, the dimension standard value and the edge weight parameter matching the dimension are multiplied to obtain the weight value, and the edge weighting is processed as multiplying the dimension standard value by the corresponding edge weight The parameter, that is, the dimension standard value is expanded according to the edge weight parameter, and the potential correlation degree between each dimension and the spine feature is opened, so that it can be identified objectively and scientifically.
S408,对所有所述加权值进行扩充,得到所述第二特征向量图。S408: Expand all the weight values to obtain the second feature vector graph.
可理解地,所述扩充为将所有所述加权值进行复制并填充,达到预设的矩阵尺寸大小,所述复制并填充为将一维的所有所述加权值整体复制直到预设的矩阵尺寸的横向长度,最后不足以复制的元素以零填充,再将该行的所有元素进行复制直到预设的矩阵尺寸的纵向长度,从而得到所述第二特征向量图。Understandably, the expansion is to copy and fill all the weights to reach a preset matrix size, and the copy and fill is to copy all the weights in one dimension as a whole to the preset matrix size. Finally, the elements that are insufficient to be copied are filled with zeros, and then all the elements of the row are copied to the longitudinal length of the preset matrix size, thereby obtaining the second eigenvector map.
本申请通过将非图像数据进行归一化及边缘权重处理并生成与脊柱变形的特征相关的第二特征向量图,能够更好的建立各个特征向量之间关系的第三特征向量图,提高识别的可靠性。This application performs normalization and edge weighting of non-image data and generates a second feature vector diagram related to the features of spine deformation, which can better establish the third feature vector diagram of the relationship between each feature vector, and improve recognition Reliability.
S50,将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图。S50: Perform edge filling of the second feature vector graph to the first feature vector graph to obtain a third feature vector graph.
可理解地,所述边缘填充指在所述第一特征向量图的基础上在数组边缘上进行填充,即将所述第二特征向量图在第一特征向量图的矩阵上方及下方进行填充至预设的尺寸大小,得到所述第三特征向量图。Understandably, the edge filling refers to filling on the edge of the array on the basis of the first eigenvector image, that is, filling the second eigenvector image above and below the matrix of the first eigenvector image to the preset value. Set the size to obtain the third feature vector diagram.
S60,将所述第三特征向量图输入训练完成的脊柱图卷积网络模型。S60: Input the third feature vector map into the trained spine map convolutional network model.
可理解地,所述脊柱图卷积网络模型为基于图卷积的网络结构进行识别并训练完成的神经网络模型,所述脊柱图卷积网络模型能够根据所述第三特征向量图中各元素的特征向量及各元素之间的关联关系进行提取脊柱频域特征,对提取到的所述脊柱频域特征进行分类识别,最后输出对应的脊柱变形人群的类别。Understandably, the spine diagram convolutional network model is a neural network model that is identified and trained based on the network structure of the diagram convolution, and the spine diagram convolutional network model can be based on the elements in the third feature vector diagram. The spine frequency domain features are extracted from the feature vector and the association relationship between each element, and the extracted spine frequency domain features are classified and recognized, and finally the category of the corresponding spine deformed crowd is output.
在一实施例中,如图7所示,所述步骤S60之前,即所述将所述第三特征向量图输入训练完成的脊柱图卷积网络模型之前,包括:In an embodiment, as shown in FIG. 7, before step S60, that is, before inputting the third feature vector image into the trained spine graph convolutional network model, the method includes:
S601,获取样本数据集;所述样本数据集包括样本数据和与样本数据一一对应的样本标签;所述样本数据为历史第三特征向量图;所述样本标签包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群和潜在脊柱驼背人群。S601. Obtain a sample data set; the sample data set includes sample data and sample labels corresponding to the sample data one-to-one; the sample data is a historical third feature vector diagram; the sample labels include scoliosis people and kyphotic people , Potential scoliosis crowd and potential spine kyphosis crowd.
可理解地,所述历史第三特征向量图为与所述样本数据关联的样本图像数据和样本非图像数据经过上述步骤S20至步骤S50处理后获得,所述样本数据集中包含多个所述样本数据,一个所述样本数据关联一个所述样本标签,所述样本标签包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群和潜在脊柱驼背人群。Understandably, the historical third feature vector graph is obtained after the sample image data and sample non-image data associated with the sample data are processed through the above-mentioned steps S20 to S50, and the sample data set contains a plurality of the samples. Data, one sample data is associated with one sample label, and the sample label includes a scoliosis crowd, a kyphosis crowd, a potential scoliosis crowd, and a potential kyphosis crowd.
S602,将所述样本数据输入含有初始参数的脊柱图卷积神经网络模型。S602: Input the sample data into a spinal diagram convolutional neural network model containing initial parameters.
可理解地,所述脊柱图卷积神经网络模型的所述初始参数可以根据需求设定,比如初始参数可以通过迁移学习方法获取其他与背部相关识别的图卷积模型的所有参数,也可以全部设置为预设的一个数值。Understandably, the initial parameters of the spine graph convolutional neural network model can be set according to requirements. For example, the initial parameters can be obtained through a transfer learning method to obtain all the parameters of other graph convolution models related to the back recognition, or all Set to a preset value.
S603,根据GCN中的频谱域方法,通过所述脊柱图卷积神经网络模型提取所述样本数据中的脊柱频域特征,获取所述脊柱图卷积神经网络模型根据所述脊柱频域特征输出的样本结果。S603, according to the spectral domain method in the GCN, extract the spine frequency domain features in the sample data through the spine map convolutional neural network model, and obtain the spine map convolutional neural network model to output according to the spine frequency domain features Sample results.
可理解地,所述GCN为Graph Convolutional Network,即为对用顶点和边建立相应关系的拓扑图进行相关空间特征的提取进行识别出存在或潜在的分类,所述频谱域方法为通过借助于拓扑图对应的拉普拉斯矩阵的特征值和特征向量来研究拓扑图的性质及对拉普拉斯矩阵的特征值和特征向量进行傅里叶变换获得的GCN分类的结果的方法,所述脊柱频域特征为通过运用所述频谱域方法转换且与脊柱变形相关的特征,所述脊柱图卷积神经网络模型根据提取的所述脊柱频域特征识别出所述样本数据的所述样本结果,所述样本结果包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。Understandably, the GCN is Graph Convolutional Network, that is, to identify the existence or potential classification by extracting relevant spatial features of a topological graph that uses vertices and edges to establish a corresponding relationship. The spectral domain method is by relying on topological The eigenvalues and eigenvectors of the Laplacian matrix corresponding to the graph are used to study the properties of the topological graph and the GCN classification result obtained by Fourier transforming the eigenvalues and eigenvectors of the Laplacian matrix. The spine The frequency domain feature is a feature that is transformed by using the frequency domain method and is related to spinal deformation, and the spine map convolutional neural network model identifies the sample result of the sample data according to the extracted frequency domain feature of the spine, The sample results include people with scoliosis, people with kyphosis, people with potential scoliosis, people with potential kyphosis, and people without spinal deformity.
S604,根据所述样本数据对应的所述样本结果和所述样本标签,确定损失值。S604: Determine a loss value according to the sample result and the sample label corresponding to the sample data.
可理解地,将所述样本结果和所述样本标签输入所述脊柱图卷积神经网络模型中损失函数,计算得到所述样本数据对应的所述损失值,所述损失函数可以根据需求设定,所述损失函数为所述样本结果和所述样本标签之间差值的对数,表明了所述样本结果和所述样本标签之间的差距。Understandably, the sample result and the sample label are input into the loss function in the spine map convolutional neural network model, and the loss value corresponding to the sample data is calculated. The loss function can be set according to requirements The loss function is the logarithm of the difference between the sample result and the sample label, and indicates the difference between the sample result and the sample label.
S605,在所述损失值达到预设的收敛条件时,将收敛之后的所述脊柱图卷积神经网络模型记录为训练完成的脊柱图卷积网络模型。S605: When the loss value reaches a preset convergence condition, record the spine diagram convolutional neural network model after convergence as a trained spine diagram convolutional network model.
可理解地,所述收敛条件可以为所述损失值小于设定阈值的条件,即在所述损失值小于设定阈值时,将收敛之后的所述脊柱图卷积神经网络模型记录为训练完成的脊柱图卷积网络模型。Understandably, the convergence condition may be a condition that the loss value is less than the set threshold, that is, when the loss value is less than the set threshold, the spine map convolutional neural network model after convergence is recorded as training completed The convolutional network model of the spine diagram.
在一实施例中,所述步骤S604之后,即所述根据所述样本数据对应的所述样本结果和所述样本标签,确定损失值之后,还包括:In an embodiment, after step S604, that is, after determining the loss value according to the sample result and the sample label corresponding to the sample data, the method further includes:
S606,在所述损失值未达到预设的收敛条件时,迭代更新所述脊柱图卷积神经网络模 型的初始参数,直至所述损失值达到所述预设的收敛条件时,将收敛之后的所述脊柱图卷积神经网络模型记录为训练完成的脊柱图卷积网络模型。S606: When the loss value does not reach the preset convergence condition, iteratively update the initial parameters of the spine map convolutional neural network model, until the loss value reaches the preset convergence condition, converge The spine map convolutional neural network model is recorded as a trained spine map convolutional network model.
可理解地,所述收敛条件也可以为所述损失值经过了10000次计算后值为很小且不会再下降的条件,即在所述损失值经过10000次计算后值为很小且不会再下降时,停止训练,将收敛之后的所述脊柱图卷积神经网络模型记录为训练完成的脊柱图卷积网络模型。Understandably, the convergence condition may also be a condition that the loss value is small and will not decrease after 10,000 calculations, that is, the loss value is small and does not fall after 10,000 calculations. When it drops again, the training is stopped, and the spine map convolutional neural network model after convergence is recorded as the spine map convolutional network model that has been trained.
如此,在所述损失值未达到预设的收敛条件时,不断更新迭代所述脊柱图卷积神经网络模型的初始参数,可以不断向准确的识别结果靠拢,让识别结果的准确率越来越高。In this way, when the loss value does not reach the preset convergence condition, the initial parameters of the iterative convolutional neural network model of the spine diagram can be continuously updated, which can continuously move closer to the accurate recognition result, so that the accuracy of the recognition result becomes more and more accurate. high.
S70,根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。S70. According to the spectral domain method in the GCN, extract the spine frequency domain features in the third feature vector map through the spine map convolution network model, and obtain the spine map convolution network model according to the spine frequency domain features The output recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes the scoliosis crowd, the kyphotic crowd, the potential scoliosis crowd, the potential kyphotic crowd, and People with non-spine deformities.
可理解地,所述GCN为Graph Convolutional Network,即为对用顶点和边建立相应关系的拓扑图进行相关空间特征的提取进行识别出存在或潜在的分类,所述频谱域方法为通过借助于拓扑图对应的拉普拉斯矩阵的特征值和特征向量来研究拓扑图的性质及对拉普拉斯矩阵的特征值和特征向量进行傅里叶变换获得的GCN分类的结果的方法,通过所述脊柱图卷积网络模型对所述待识别目标的背部照片进行识别,确定出所述待识别目标所处所述脊柱变形人群的类别,能够对潜在人群进行提醒,确定之后可以针对所述脊柱变形人群的类别输出对应的提醒提示或者预防提示。Understandably, the GCN is Graph Convolutional Network, that is, to identify the existence or potential classification by extracting relevant spatial features of a topological graph that uses vertices and edges to establish a corresponding relationship. The spectral domain method is by means of topology The eigenvalues and eigenvectors of the Laplacian matrix corresponding to the graph are used to study the properties of the topological graph and the method of performing the Fourier transform on the eigenvalues and eigenvectors of the Laplacian matrix to obtain the result of GCN classification. The spine map convolutional network model recognizes the back photos of the target to be recognized, determines the category of the spine deformed crowd where the target to be recognized is located, can remind potential crowds, and can target the spine deformed after determination The corresponding reminder or prevention reminder is output for the category of the crowd.
本申请通过接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;将所述图像数据输入背部区域识别模型,获取所述背部区域识别模型截取出的待识别背部区域图像;对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果。This application obtains the image data and non-image data associated with the unique code corresponding to the target to be recognized by receiving the target recognition instruction; inputs the image data into the back area recognition model, and obtains the back to be recognized cut out by the back area recognition model Regional image; perform image enhancement processing on the image of the back region to be recognized to obtain an enhanced image of the region to be recognized; input the enhanced image of the region to be recognized into a spine recognition model, and extract the enhanced image of the region to be recognized through the spine recognition model The spine feature in the spine, acquiring the first feature vector diagram output by the spine recognition model according to the spine feature, and inputting the non-image data into the data standardization model, and classifying the non-image data through the data standardization model Unification and edge weight processing to obtain the second eigenvector image; edge-filling the second eigenvector image to the first eigenvector image to obtain the third eigenvector image; according to the spectral domain method in GCN, pass The spine map convolution network model extracts the spine frequency domain features in the third feature vector map, and obtains the recognition result output by the spine map convolution network model according to the spine frequency domain features.
本申请实现了通过获取与待识别目标关联的图像数据和非图像数据;截取出所述图像数据中背部区域的待识别背部区域图像;对所述待识别背部区域图像进行图像增强处理得到待识别区域增强图像;通过脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获得第一特征向量图,同时通过数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;根据GCN中的频谱域方法,通过脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果,如此,本申请实现了根据待识别目标的拍摄的背部的图像和相关的非图像信息,通过GCN中的频谱域方法自动识别出待识别目标对应的脊柱变形人群的类别(包括潜在的人群侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群),能够快速地、准确地识别出待识别目标对应脊柱变形人群的类别,对潜在的人群进行提醒,起到预防作用,因此,提高了脊柱变形人群识别的准确率和可靠性,大大降低了识别成本,并对潜在的人群起到提醒作用。This application realizes that by acquiring the image data and non-image data associated with the target to be recognized; intercepting the image of the back area to be recognized in the back area of the image data; performing image enhancement processing on the image of the back area to be recognized to obtain the to be recognized Region-enhanced image; extract the features of the spine in the enhanced image of the region to be identified through the spine recognition model to obtain the first feature vector diagram, and at the same time normalize the non-image data and process the edge weights through the data standardization model to obtain the first Second feature vector graph; edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph; according to the spectral domain method in GCN, the spine graph convolutional network model is used to extract the The frequency domain features of the spine in the third feature vector diagram are obtained, and the recognition results output by the convolutional network model of the spine diagram according to the frequency domain features of the spine are obtained. Related non-image information, through the spectrum domain method in the GCN, automatically identify the category of the spine deformed population corresponding to the target to be identified (including the potential crowd scoliosis crowd, spine kyphosis crowd, potential scoliosis crowd, potential spine kyphosis crowd and potential scoliosis crowd). People with non-spine deformities) can quickly and accurately identify the categories of people with spine deformities corresponding to the target to be identified, and remind potential people to prevent them. Therefore, the accuracy and reliability of the recognition of people with spine deformities are improved. , Which greatly reduces the cost of identification and serves as a reminder to potential crowds.
在一实施例中,提供一种脊柱变形人群识别装置,该脊柱变形人群识别装置与上述实施例中脊柱变形人群识别方法一一对应。如图8所示,该脊柱变形人群识别装置包括接收 模块11、识别模块12、增强模块13、获取模块14、填充模块15、输入模块16和输出模块17。各功能模块详细说明如下:In one embodiment, a device for identifying people with deformed spine is provided, and the device for identifying people with deformed spine corresponds one-to-one with the method for identifying people with deformed spine in the above-mentioned embodiment. As shown in FIG. 8, the device for identifying people with deformed spine includes a receiving module 11, a recognition module 12, an enhancement module 13, an acquisition module 14, a filling module 15, an input module 16 and an output module 17. The detailed description of each functional module is as follows:
接收模块11,用于接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;The receiving module 11 is configured to receive a target recognition instruction, and obtain image data and non-image data associated with the unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is the target to be recognized Related information;
识别模块12,用于将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;The recognition module 12 is configured to input the image data into a back area recognition model, and perform back area recognition on the image data through the back area recognition model, and obtain the back area image to be recognized cut out by the back area recognition model ; The back region recognition model is a deep convolutional neural network model based on the YOLO model;
增强模块13,用于对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;The enhancement module 13 is configured to perform image enhancement processing on the image of the back area to be identified to obtain an enhanced image of the area to be identified;
获取模块14,用于将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The acquiring module 14 is configured to input the enhanced image of the region to be recognized into a spine recognition model, extract the features of the spine in the enhanced image of the region to be identified through the spine recognition model, and obtain the spine recognition model and output according to the spine features At the same time, the non-image data is input into a data standardization model, and the non-image data is normalized and edge weighted through the data standardization model to obtain a second feature vector diagram;
填充模块15,用于将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;The filling module 15 is used for edge filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
输入模块16,用于将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;The input module 16 is configured to input the third feature vector map into the trained spine map convolutional network model;
输出模块17,用于根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。The output module 17 is configured to extract the spine frequency domain features in the third feature vector diagram through the spine map convolutional network model according to the spectral domain method in the GCN, and obtain the spine map convolutional network model according to the spine map convolutional network model. The recognition result of the spine frequency domain feature output; the recognition result characterizes the category of the spine deformed population of the target to be recognized, and the category of the spine deformed population includes the scoliosis population, the kyphotic population, the potential scoliosis population, and the potential scoliosis population. People with kyphosis and non-spine deformities.
关于脊柱变形人群识别装置的具体限定可以参见上文中对于脊柱变形人群识别方法的限定,在此不再赘述。上述脊柱变形人群识别装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。For the specific definition of the device for identifying people with deformed spine, please refer to the above definition of the method for identifying people with deformed spine, which will not be repeated here. The various modules in the device for identifying people with deformed spine can be implemented in whole or in part by software, hardware, and a combination thereof. The above-mentioned modules may be embedded in the form of hardware or independent of the processor in the computer equipment, or may be stored in the memory of the computer equipment in the form of software, so that the processor can call and execute the operations corresponding to the above-mentioned modules.
在一个实施例中,提供了一种计算机设备,该计算机设备可以是服务器,其内部结构图可以如图9所示。该计算机设备包括通过系统总线连接的处理器、存储器、网络接口和数据库。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统、计算机程序和数据库。该内存储器为非易失性存储介质中的操作系统和计算机程序的运行提供环境。该计算机设备的网络接口用于与外部的终端通过网络连接通信。该计算机程序被处理器执行时以实现一种脊柱变形人群识别方法。In one embodiment, a computer device is provided. The computer device may be a server, and its internal structure diagram may be as shown in FIG. 9. The computer equipment includes a processor, a memory, a network interface, and a database connected through a system bus. Among them, the processor of the computer device is used to provide calculation and control capabilities. The memory of the computer device includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, a computer program and a database. The internal memory provides an environment for the operation of the operating system and computer programs in the non-volatile storage medium. The network interface of the computer device is used to communicate with an external terminal through a network connection. The computer program is executed by the processor to realize a method for identifying people with deformed spine.
在一个实施例中,提供了一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,处理器执行计算机程序时实现上述实施例中脊柱变形人群识别方法。In one embodiment, a computer device is provided, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor. The processor executes the computer program to implement the method for identifying people with deformed spine in the above-mentioned embodiment. .
在一个实施例中,提供了一种计算机可读存储介质,其上存储有计算机程序,计算机程序被处理器执行时实现上述实施例中脊柱变形人群识别方法。可选的,本申请涉及的存储介质如计算机可读存储介质可以是非易失性的,也可以是易失性的。In one embodiment, a computer-readable storage medium is provided, on which a computer program is stored, and when the computer program is executed by a processor, the method for identifying people with deformed spine in the above-mentioned embodiment is realized. Optionally, the storage medium involved in this application, such as a computer-readable storage medium, may be non-volatile or volatile.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一非易失性计算机可读取存储介质中,该计算机程序在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可 包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。A person of ordinary skill in the art can understand that all or part of the processes in the above-mentioned embodiment methods can be implemented by instructing relevant hardware through a computer program. The computer program can be stored in a non-volatile computer readable storage. In the medium, when the computer program is executed, it may include the processes of the above-mentioned method embodiments. Wherein, any reference to memory, storage, database or other media used in the embodiments provided in this application may include non-volatile and/or volatile memory. Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory may include random access memory (RAM) or external cache memory. As an illustration and not a limitation, RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。Those skilled in the art can clearly understand that, for the convenience and conciseness of description, only the division of the above functional units and modules is used as an example. In practical applications, the above functions can be allocated to different functional units and modules as needed. Module completion, that is, the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that they can still implement the foregoing The technical solutions recorded in the examples are modified, or some of the technical features are equivalently replaced; these modifications or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the application, and should be included in Within the scope of protection of this application.

Claims (20)

  1. 一种脊柱变形人群识别方法,其中,包括:A method for identifying people with deformed spine, which includes:
    接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;Receiving a target recognition instruction, and obtaining image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is information related to the target to be recognized;
    将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;Input the image data into a back region recognition model, and perform back region recognition on the image data through the back region recognition model, and obtain the back region image to be recognized cut out by the back region recognition model; the back region recognition The model is a deep convolutional neural network model based on the YOLO model;
    对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;Performing image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
    将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained At the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
    将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;Edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
    将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;Input the third feature vector map into the trained spine map convolutional network model;
    根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。According to the spectral domain method in GCN, the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained. Recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
  2. 如权利要求1所述的脊柱变形人群识别方法,其中,所述通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像,包括:The method for recognizing people with deformed spine according to claim 1, wherein the image data is recognized by the back area recognition model to obtain the back area image to be recognized cut out by the back area recognition model ,include:
    将目标背部背面图像输入所述背部区域识别模型中的背部背面区域识别模型,同时将目标背部侧面图像输入所述背部区域识别模型中的背部侧面区域识别模型;所述图像数据包括所述目标背部背面图像和所述目标背部侧面图像;Input the back image of the target back into the back back area recognition model in the back area recognition model, and input the back side image of the target into the back side area recognition model in the back area recognition model; the image data includes the back of the target The back image and the back side image of the target;
    根据YOLO算法,通过所述背部背面区域识别模型进行识别,截取出只含有所述待识别目标的背部背面的待识别背部背面区域图像,同时通过所述背部侧面区域识别模型进行识别,截取出只含有所述待识别目标的背部侧面的待识别背部侧面区域图像;According to the YOLO algorithm, the recognition is performed through the back area recognition model, and the image of the back back area to be recognized that contains only the back back of the target to be recognized is cut out, and the back side area recognition model is used for recognition. An image of the back side area to be recognized containing the back side of the target to be recognized;
    将所述待识别背部背面区域图像和所述待识别背部侧面区域图像确定为所述待识别背部区域图像。The image of the back area to be recognized and the image of the back side area to be recognized are determined as the image of the back area to be recognized.
  3. 如权利要求1所述的脊柱变形人群识别方法,其中,所述对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像,包括:The method for identifying people with deformed spine as claimed in claim 1, wherein said performing image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified comprises:
    对所述待识别背部区域图像中的所述待识别背部背面区域图像进行灰度化处理,得到背部背面灰度图像,同时对所述待识别背部区域图像中的所述待识别背部侧面区域图像进行灰度化处理,得到背部侧面灰度图像;Perform gray-scale processing on the image of the back area to be recognized in the image of the back area to be recognized to obtain a gray-scale image of the back back, and at the same time perform the gray-scale processing on the image of the back side area to be recognized in the back area image to be recognized Perform gray-scale processing to obtain a gray-scale image of the back side;
    对所述背部背面灰度图像进行图像去噪音及边缘增强处理,得到背部背面增强图像,同时对所述背部侧面灰度图像进行图像去噪音及边缘增强处理,得到背部侧面增强图像;Performing image denoising and edge enhancement processing on the back and back grayscale image to obtain a back back enhanced image, and simultaneously performing image denoising and edge enhancement processing on the back side grayscale image to obtain a back side enhanced image;
    将所述背部背面增强图像和所述背部侧面增强图像确定为所述待识别区域增强图像。The back back enhanced image and the back side enhanced image are determined as the enhanced image of the region to be recognized.
  4. 如权利要求1所述的脊柱变形人群识别方法,其中,所述脊柱特征包括侧弯特征和驼背特征;The method for identifying people with deformed spine according to claim 1, wherein the spine features include scoliosis features and kyphosis features;
    所述通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,包括:The extracting the spine features in the enhanced image of the region to be recognized through the spine recognition model, and obtaining the first feature vector diagram output by the spine recognition model according to the spine features, includes:
    通过侧弯识别模型对所述背部背面增强图像进行所述侧弯特征提取,同时通过驼背识别模型对所述背部侧面增强图像进行所述驼背特征提取;所述脊柱识别模型包括所述侧弯识别模型和所述驼背识别模型;The scoliosis feature extraction is performed on the back enhanced image through a scoliosis recognition model, and the kyphosis feature extraction is performed on the back side enhanced image through a kyphosis recognition model; the spine recognition model includes the scoliosis recognition Model and said humpback recognition model;
    获取所述侧弯识别模型根据所述侧弯特征输出的侧弯特征向量图,同时获取所述驼背识别模型根据所述驼背特征输出的驼背特征向量图;Acquiring a side curve feature vector diagram output by the side curve recognition model according to the side curve feature, and simultaneously acquiring a hunchback feature vector diagram output by the hunchback recognition model according to the hunchback feature;
    将所述侧弯特征向量图与所述驼背特征向量图进行拼接,得到所述第一特征向量图。The side curve feature vector graph and the humpback feature vector graph are spliced together to obtain the first feature vector graph.
  5. 如权利要求1所述的脊柱变形人群识别方法,其中,所述通过数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图,包括:The method for identifying people with deformed spine according to claim 1, wherein said normalizing and edge weighting said non-image data through a data standardization model to obtain a second feature vector map comprises:
    获取所述非图像数据中的各个维度和与各个所述维度对应的维度数据;Acquiring each dimension in the non-image data and dimension data corresponding to each of the dimensions;
    获取与各个所述维度匹配的归一化规则和边缘权重参数;Obtaining normalization rules and edge weight parameters that match each of the dimensions;
    根据与各个所述维度匹配的归一化规则,对所有所述维度数据进行归一化处理,得到与各个所述维度对应的维度标准值;Performing normalization processing on all the dimensional data according to the normalization rule matching each of the dimensions to obtain the dimension standard value corresponding to each of the dimensions;
    根据与各个所述维度匹配的边缘权重参数,对所有与所述维度标准值进行边缘加权处理,得到与各个所述维度对应的加权值;Performing edge weighting processing on all standard values of the dimensions according to the edge weight parameters matched with each of the dimensions to obtain weighted values corresponding to each of the dimensions;
    对所有所述加权值进行扩充,得到所述第二特征向量图。All the weighted values are expanded to obtain the second feature vector graph.
  6. 如权利要求5所述的脊柱变形人群识别方法,其中,所述非图像数据中的维度包括目标性别、目标年龄、目标职业和目标信息。5. The method for identifying people with deformed spine, wherein the dimensions in the non-image data include target gender, target age, target occupation, and target information.
  7. 如权利要求1所述的脊柱变形人群识别方法,其中,所述将所述第三特征向量图输入训练完成的脊柱图卷积网络模型之前,包括:The method for identifying people with deformed spine according to claim 1, wherein, before inputting the third feature vector map into the trained spine map convolutional network model, the method comprises:
    获取样本数据集;所述样本数据集包括样本数据和与样本数据一一对应的样本标签;所述样本数据为历史第三特征向量图;所述样本标签包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群和潜在脊柱驼背人群;Obtain a sample data set; the sample data set includes sample data and sample labels corresponding to the sample data one-to-one; the sample data is a historical third feature vector diagram; the sample labels include scoliosis crowds, kyphotic crowds, and potential People with scoliosis and people with potential kyphosis;
    将所述样本数据输入含有初始参数的脊柱图卷积神经网络模型;Inputting the sample data into a spine diagram convolutional neural network model containing initial parameters;
    根据GCN中的频谱域方法,通过所述脊柱图卷积神经网络模型提取所述样本数据中的脊柱频域特征,获取所述脊柱图卷积神经网络模型根据所述脊柱频域特征输出的样本结果;According to the spectral domain method in GCN, the spine frequency domain features in the sample data are extracted through the spine map convolutional neural network model, and the samples output by the spine map convolutional neural network model according to the spine frequency domain features are obtained result;
    根据所述样本数据对应的所述样本结果和所述样本标签,确定损失值;Determine a loss value according to the sample result and the sample label corresponding to the sample data;
    在所述损失值达到预设的收敛条件时,将收敛之后的所述脊柱图卷积神经网络模型记录为训练完成的脊柱图卷积网络模型。When the loss value reaches a preset convergence condition, the spine diagram convolutional neural network model after convergence is recorded as the trained spine diagram convolutional network model.
  8. 一种脊柱变形人群识别装置,其中,包括:A device for identifying people with deformed spine, which includes:
    接收模块,用于接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;The receiving module is used to receive a target recognition instruction, and obtain image data and non-image data associated with the unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is related to the target to be recognized Information;
    识别模块,用于将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;A recognition module, configured to input the image data into a back area recognition model, perform back area recognition on the image data through the back area recognition model, and obtain an image of the back area to be recognized that is intercepted by the back area recognition model; The back region recognition model is a deep convolutional neural network model based on the YOLO model;
    增强模块,用于对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;An enhancement module, configured to perform image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
    获取模块,用于将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The acquiring module is configured to input the enhanced image of the region to be recognized into a spine recognition model, extract the features of the spine in the enhanced image of the region to be identified through the spine recognition model, and obtain the output of the spine recognition model according to the features of the spine A first feature vector map, and at the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
    填充模块,用于将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;A filling module, configured to perform edge filling of the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
    输入模块,用于将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;An input module, configured to input the third feature vector graph into the trained spine graph convolutional network model;
    输出模块,用于根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变 形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。The output module is used to extract the frequency domain features of the spine in the third feature vector diagram through the spine map convolutional network model according to the spectral domain method in the GCN, and obtain the spine map convolutional network model according to the spine Frequency domain feature output recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be recognized, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, and potential spine crowd People with hunchbacks and non-spine deformities.
  9. 一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其中,所述处理器执行所述计算机程序时实现以下步骤:A computer device includes a memory, a processor, and a computer program that is stored in the memory and can run on the processor, wherein the processor implements the following steps when the processor executes the computer program:
    接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;Receiving a target recognition instruction, and obtaining image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is information related to the target to be recognized;
    将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;Input the image data into a back region recognition model, and perform back region recognition on the image data through the back region recognition model, and obtain the back region image to be recognized cut out by the back region recognition model; the back region recognition The model is a deep convolutional neural network model based on the YOLO model;
    对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;Performing image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
    将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained At the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
    将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;Edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
    将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;Input the third feature vector map into the trained spine map convolutional network model;
    根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。According to the spectral domain method in GCN, the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained. Recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
  10. 如权利要求9所述的计算机设备,其中,所述通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像时,具体实现以下步骤:The computer device according to claim 9, wherein, when the image data is recognized by the back area recognition model to recognize the back area, and the back area image to be recognized extracted by the back area recognition model is obtained, specifically Implement the following steps:
    将目标背部背面图像输入所述背部区域识别模型中的背部背面区域识别模型,同时将目标背部侧面图像输入所述背部区域识别模型中的背部侧面区域识别模型;所述图像数据包括所述目标背部背面图像和所述目标背部侧面图像;Input the back image of the target back into the back back area recognition model in the back area recognition model, and input the back side image of the target into the back side area recognition model in the back area recognition model; the image data includes the back of the target The back image and the back side image of the target;
    根据YOLO算法,通过所述背部背面区域识别模型进行识别,截取出只含有所述待识别目标的背部背面的待识别背部背面区域图像,同时通过所述背部侧面区域识别模型进行识别,截取出只含有所述待识别目标的背部侧面的待识别背部侧面区域图像;According to the YOLO algorithm, the recognition is performed by the back back area recognition model, and the back back area image to be recognized that contains only the back back of the target to be recognized is cut out, and at the same time, the back side area recognition model is used for recognition, and only the back back area recognition model is cut out. An image of the back side area to be recognized containing the back side of the target to be recognized;
    将所述待识别背部背面区域图像和所述待识别背部侧面区域图像确定为所述待识别背部区域图像。The image of the back area to be recognized and the image of the back side area to be recognized are determined as the image of the back area to be recognized.
  11. 如权利要求9所述的计算机设备,其中,所述对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像时,具体实现以下步骤:9. The computer device according to claim 9, wherein when the image enhancement processing is performed on the image of the back region to be recognized to obtain the enhanced image of the region to be recognized, the following steps are specifically implemented:
    对所述待识别背部区域图像中的所述待识别背部背面区域图像进行灰度化处理,得到背部背面灰度图像,同时对所述待识别背部区域图像中的所述待识别背部侧面区域图像进行灰度化处理,得到背部侧面灰度图像;Perform gray-scale processing on the image of the back area to be recognized in the image of the back area to be recognized to obtain a gray-scale image of the back back, and at the same time perform the gray-scale processing on the image of the back side area to be recognized in the back area image to be recognized Perform gray-scale processing to obtain a gray-scale image of the back side;
    对所述背部背面灰度图像进行图像去噪音及边缘增强处理,得到背部背面增强图像,同时对所述背部侧面灰度图像进行图像去噪音及边缘增强处理,得到背部侧面增强图像;Performing image denoising and edge enhancement processing on the back and back grayscale image to obtain a back back enhanced image, and simultaneously performing image denoising and edge enhancement processing on the back side grayscale image to obtain a back side enhanced image;
    将所述背部背面增强图像和所述背部侧面增强图像确定为所述待识别区域增强图像。The back enhanced image and the back side enhanced image are determined as the enhanced image of the region to be recognized.
  12. 如权利要求9所述的计算机设备,其中,所述脊柱特征包括侧弯特征和驼背特征;9. The computer device of claim 9, wherein the spine features include side curvature features and kyphosis features;
    所述通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图时,具体实现以下步骤:When extracting the spine features in the enhanced image of the region to be recognized through the spine recognition model, and obtaining the first feature vector map output by the spine recognition model according to the spine features, the following steps are specifically implemented:
    通过侧弯识别模型对所述背部背面增强图像进行所述侧弯特征提取,同时通过驼背识别模型对所述背部侧面增强图像进行所述驼背特征提取;所述脊柱识别模型包括所述侧弯 识别模型和所述驼背识别模型;The scoliosis feature extraction is performed on the back enhanced image through a scoliosis recognition model, and the kyphosis feature extraction is performed on the back side enhanced image through a kyphosis recognition model; the spine recognition model includes the scoliosis recognition Model and said humpback recognition model;
    获取所述侧弯识别模型根据所述侧弯特征输出的侧弯特征向量图,同时获取所述驼背识别模型根据所述驼背特征输出的驼背特征向量图;Acquiring a side curve feature vector diagram output by the side curve recognition model according to the side curve feature, and simultaneously acquiring a hunchback feature vector diagram output by the hunchback recognition model according to the hunchback feature;
    将所述侧弯特征向量图与所述驼背特征向量图进行拼接,得到所述第一特征向量图。The side curve feature vector graph and the humpback feature vector graph are spliced together to obtain the first feature vector graph.
  13. 如权利要求9所述的计算机设备,其中,所述通过数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图时,具体实现以下步骤:9. The computer device according to claim 9, wherein, when the non-image data is normalized and edge weighted by the data standardization model to obtain the second feature vector map, the following steps are specifically implemented:
    获取所述非图像数据中的各个维度和与各个所述维度对应的维度数据;Acquiring each dimension in the non-image data and dimension data corresponding to each of the dimensions;
    获取与各个所述维度匹配的归一化规则和边缘权重参数;Obtaining normalization rules and edge weight parameters that match each of the dimensions;
    根据与各个所述维度匹配的归一化规则,对所有所述维度数据进行归一化处理,得到与各个所述维度对应的维度标准值;Performing normalization processing on all the dimensional data according to the normalization rule matching each of the dimensions to obtain the dimension standard value corresponding to each of the dimensions;
    根据与各个所述维度匹配的边缘权重参数,对所有与所述维度标准值进行边缘加权处理,得到与各个所述维度对应的加权值;Performing edge weighting processing on all standard values of the dimensions according to the edge weight parameters matched with each of the dimensions to obtain weighted values corresponding to each of the dimensions;
    对所有所述加权值进行扩充,得到所述第二特征向量图。All the weighted values are expanded to obtain the second feature vector graph.
  14. 如权利要求9所述的计算机设备,其中,所述将所述第三特征向量图输入训练完成的脊柱图卷积网络模型之前,所述处理器执行所述计算机程序时还实现以下步骤:9. The computer device according to claim 9, wherein before the input of the third feature vector map into the trained spine map convolutional network model, the processor further implements the following steps when executing the computer program:
    获取样本数据集;所述样本数据集包括样本数据和与样本数据一一对应的样本标签;所述样本数据为历史第三特征向量图;所述样本标签包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群和潜在脊柱驼背人群;Obtain a sample data set; the sample data set includes sample data and sample labels corresponding to the sample data one-to-one; the sample data is a historical third feature vector diagram; the sample labels include scoliosis crowds, kyphotic crowds, and potential People with scoliosis and people with potential kyphosis;
    将所述样本数据输入含有初始参数的脊柱图卷积神经网络模型;Inputting the sample data into a spine diagram convolutional neural network model containing initial parameters;
    根据GCN中的频谱域方法,通过所述脊柱图卷积神经网络模型提取所述样本数据中的脊柱频域特征,获取所述脊柱图卷积神经网络模型根据所述脊柱频域特征输出的样本结果;According to the spectral domain method in GCN, the spine frequency domain features in the sample data are extracted through the spine map convolutional neural network model, and the samples output by the spine map convolutional neural network model according to the spine frequency domain features are obtained result;
    根据所述样本数据对应的所述样本结果和所述样本标签,确定损失值;Determine a loss value according to the sample result and the sample label corresponding to the sample data;
    在所述损失值达到预设的收敛条件时,将收敛之后的所述脊柱图卷积神经网络模型记录为训练完成的脊柱图卷积网络模型。When the loss value reaches a preset convergence condition, the spine diagram convolutional neural network model after convergence is recorded as the trained spine diagram convolutional network model.
  15. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其中,所述计算机程序被处理器执行时实现以下步骤:A computer-readable storage medium storing a computer program, wherein the computer program is executed by a processor to implement the following steps:
    接收目标识别指令,获取与待识别目标对应的唯一码关联的图像数据和非图像数据;所述图像数据为与背部相关的图像;所述非图像数据为与待识别目标相关的信息;Receiving a target recognition instruction, and obtaining image data and non-image data associated with a unique code corresponding to the target to be recognized; the image data is an image related to the back; the non-image data is information related to the target to be recognized;
    将所述图像数据输入背部区域识别模型,通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部区域图像;所述背部区域识别模型为基于YOLO模型搭架的深度卷积神经网络模型;Input the image data into a back region recognition model, and perform back region recognition on the image data through the back region recognition model, and obtain the back region image to be recognized cut out by the back region recognition model; the back region recognition The model is a deep convolutional neural network model based on the YOLO model;
    对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像;Performing image enhancement processing on the image of the back region to be identified to obtain an enhanced image of the region to be identified;
    将所述待识别区域增强图像输入脊柱识别模型,通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图,同时将所述非图像数据输入数据标准化模型,通过所述数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图;The enhanced image of the to-be-recognized area is input into a spine recognition model, the spine features in the enhanced image of the to-be-recognized area are extracted through the spine recognition model, and a first feature vector image output by the spine recognition model according to the spine feature is obtained At the same time input the non-image data into a data standardization model, and perform normalization and edge weight processing on the non-image data through the data standardization model to obtain a second feature vector map;
    将所述第二特征向量图进行边缘填充至所述第一特征向量图,得到第三特征向量图;Edge-filling the second feature vector graph to the first feature vector graph to obtain a third feature vector graph;
    将所述第三特征向量图输入训练完成的脊柱图卷积网络模型;Input the third feature vector map into the trained spine map convolutional network model;
    根据GCN中的频谱域方法,通过所述脊柱图卷积网络模型提取所述第三特征向量图中的脊柱频域特征,获取所述脊柱图卷积网络模型根据所述脊柱频域特征输出的识别结果;所述识别结果表征了所述待识别目标的脊柱变形人群的类别,所述脊柱变形人群的类别包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群、潜在脊柱驼背人群和非脊柱变形人群。According to the spectral domain method in GCN, the spine frequency domain features in the third feature vector diagram are extracted through the spine map convolutional network model, and the output of the spine map convolution network model according to the spine frequency domain features is obtained. Recognition result; the recognition result characterizes the category of the spine deformed crowd of the target to be identified, and the category of the spine deformed crowd includes scoliosis crowd, kyphotic crowd, potential scoliosis crowd, potential kyphotic crowd, and non-spine Deformed crowd.
  16. 如权利要求15所述的计算机可读存储介质,其中,所述通过所述背部区域识别模型对所述图像数据进行背部区域的识别,获取所述背部区域识别模型截取出的待识别背部 区域图像时,具体实现以下步骤:The computer-readable storage medium according to claim 15, wherein the back area recognition is performed on the image data through the back area recognition model to obtain the back area image to be recognized cut by the back area recognition model When, specifically implement the following steps:
    将目标背部背面图像输入所述背部区域识别模型中的背部背面区域识别模型,同时将目标背部侧面图像输入所述背部区域识别模型中的背部侧面区域识别模型;所述图像数据包括所述目标背部背面图像和所述目标背部侧面图像;Input the back image of the target back into the back back area recognition model in the back area recognition model, and input the back side image of the target into the back side area recognition model in the back area recognition model; the image data includes the back of the target The back image and the back side image of the target;
    根据YOLO算法,通过所述背部背面区域识别模型进行识别,截取出只含有所述待识别目标的背部背面的待识别背部背面区域图像,同时通过所述背部侧面区域识别模型进行识别,截取出只含有所述待识别目标的背部侧面的待识别背部侧面区域图像;According to the YOLO algorithm, the recognition is performed by the back back area recognition model, and the back back area image to be recognized that contains only the back back of the target to be recognized is cut out, and the back side area recognition model is used for recognition at the same time, and only An image of the back side area to be recognized containing the back side of the target to be recognized;
    将所述待识别背部背面区域图像和所述待识别背部侧面区域图像确定为所述待识别背部区域图像。The image of the back area to be recognized and the image of the back side area to be recognized are determined as the image of the back area to be recognized.
  17. 如权利要求15所述的计算机可读存储介质,其中,所述对所述待识别背部区域图像进行图像增强处理,得到待识别区域增强图像时,具体实现以下步骤:15. The computer-readable storage medium according to claim 15, wherein when the image enhancement processing is performed on the image of the back region to be recognized to obtain the enhanced image of the region to be recognized, the following steps are specifically implemented:
    对所述待识别背部区域图像中的所述待识别背部背面区域图像进行灰度化处理,得到背部背面灰度图像,同时对所述待识别背部区域图像中的所述待识别背部侧面区域图像进行灰度化处理,得到背部侧面灰度图像;Perform gray-scale processing on the image of the back area to be recognized in the image of the back area to be recognized to obtain a gray-scale image of the back back, and at the same time perform the gray-scale processing on the image of the back side area to be recognized in the back area image Perform gray-scale processing to obtain a gray-scale image of the back side;
    对所述背部背面灰度图像进行图像去噪音及边缘增强处理,得到背部背面增强图像,同时对所述背部侧面灰度图像进行图像去噪音及边缘增强处理,得到背部侧面增强图像;Performing image denoising and edge enhancement processing on the back and back grayscale image to obtain a back back enhanced image, and simultaneously performing image denoising and edge enhancement processing on the back side grayscale image to obtain a back side enhanced image;
    将所述背部背面增强图像和所述背部侧面增强图像确定为所述待识别区域增强图像。The back enhanced image and the back side enhanced image are determined as the enhanced image of the region to be recognized.
  18. 如权利要求15所述的计算机可读存储介质,其中,所述脊柱特征包括侧弯特征和驼背特征;15. The computer-readable storage medium of claim 15, wherein the spinal features include lateral curvature and kyphosis;
    所述通过所述脊柱识别模型提取所述待识别区域增强图像中的脊柱特征,获取所述脊柱识别模型根据所述脊柱特征输出的第一特征向量图时,具体实现以下步骤:When extracting the spine features in the enhanced image of the region to be recognized through the spine recognition model, and obtaining the first feature vector map output by the spine recognition model according to the spine features, the following steps are specifically implemented:
    通过侧弯识别模型对所述背部背面增强图像进行所述侧弯特征提取,同时通过驼背识别模型对所述背部侧面增强图像进行所述驼背特征提取;所述脊柱识别模型包括所述侧弯识别模型和所述驼背识别模型;The scoliosis feature extraction is performed on the back enhanced image through a scoliosis recognition model, and the kyphosis feature extraction is performed on the back side enhanced image through a kyphosis recognition model; the spine recognition model includes the scoliosis recognition Model and said humpback recognition model;
    获取所述侧弯识别模型根据所述侧弯特征输出的侧弯特征向量图,同时获取所述驼背识别模型根据所述驼背特征输出的驼背特征向量图;Obtaining a side curve feature vector diagram output by the side curve recognition model according to the side curve feature, and simultaneously obtaining a hunchback feature vector diagram output by the hunchback recognition model according to the hunchback feature;
    将所述侧弯特征向量图与所述驼背特征向量图进行拼接,得到所述第一特征向量图。The side curve feature vector graph and the humpback feature vector graph are spliced together to obtain the first feature vector graph.
  19. 如权利要求15所述的计算机可读存储介质,其中,所述通过数据标准化模型对所述非图像数据进行归一化及边缘权重处理,获得第二特征向量图时,具体实现以下步骤:15. The computer-readable storage medium according to claim 15, wherein when the non-image data is normalized and edge weighted by the data normalization model to obtain the second feature vector map, the following steps are specifically implemented:
    获取所述非图像数据中的各个维度和与各个所述维度对应的维度数据;Acquiring each dimension in the non-image data and dimension data corresponding to each of the dimensions;
    获取与各个所述维度匹配的归一化规则和边缘权重参数;Obtaining normalization rules and edge weight parameters that match each of the dimensions;
    根据与各个所述维度匹配的归一化规则,对所有所述维度数据进行归一化处理,得到与各个所述维度对应的维度标准值;Performing a normalization process on all the dimension data according to the normalization rule matching each dimension to obtain the dimension standard value corresponding to each dimension;
    根据与各个所述维度匹配的边缘权重参数,对所有与所述维度标准值进行边缘加权处理,得到与各个所述维度对应的加权值;Performing edge weighting processing on all standard values of the dimensions according to the edge weight parameters matched with each of the dimensions to obtain weighted values corresponding to each of the dimensions;
    对所有所述加权值进行扩充,得到所述第二特征向量图。All the weighted values are expanded to obtain the second feature vector graph.
  20. 如权利要求15所述的计算机可读存储介质,其中,所述将所述第三特征向量图输入训练完成的脊柱图卷积网络模型之前,所述计算机程序被处理器执行时还实现以下步骤:The computer-readable storage medium according to claim 15, wherein before the input of the third feature vector map into the trained spine map convolutional network model, the computer program further implements the following steps when being executed by the processor :
    获取样本数据集;所述样本数据集包括样本数据和与样本数据一一对应的样本标签;所述样本数据为历史第三特征向量图;所述样本标签包括侧弯人群、脊柱驼背人群、潜在脊柱侧弯人群和潜在脊柱驼背人群;Obtain a sample data set; the sample data set includes sample data and sample labels corresponding to the sample data one-to-one; the sample data is a historical third feature vector diagram; the sample labels include scoliosis crowds, kyphotic crowds, and potential People with scoliosis and people with potential kyphosis;
    将所述样本数据输入含有初始参数的脊柱图卷积神经网络模型;Inputting the sample data into a spine diagram convolutional neural network model containing initial parameters;
    根据GCN中的频谱域方法,通过所述脊柱图卷积神经网络模型提取所述样本数据中的脊柱频域特征,获取所述脊柱图卷积神经网络模型根据所述脊柱频域特征输出的样本结果;According to the spectral domain method in GCN, the spine frequency domain features in the sample data are extracted through the spine map convolutional neural network model, and the samples output by the spine map convolutional neural network model according to the spine frequency domain features are obtained result;
    根据所述样本数据对应的所述样本结果和所述样本标签,确定损失值;Determine a loss value according to the sample result and the sample label corresponding to the sample data;
    在所述损失值达到预设的收敛条件时,将收敛之后的所述脊柱图卷积神经网络模型记录为训练完成的脊柱图卷积网络模型。When the loss value reaches a preset convergence condition, the spine map convolutional neural network model after convergence is recorded as the trained spine map convolutional network model.
PCT/CN2020/099253 2020-06-08 2020-06-30 Method, apparatus, computer device, and storage medium for identifying persons having deformed spinal columns WO2021114623A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010513066.6A CN111666890B (en) 2020-06-08 2020-06-08 Spine deformation crowd identification method and device, computer equipment and storage medium
CN202010513066.6 2020-06-08

Publications (1)

Publication Number Publication Date
WO2021114623A1 true WO2021114623A1 (en) 2021-06-17

Family

ID=72385746

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/099253 WO2021114623A1 (en) 2020-06-08 2020-06-30 Method, apparatus, computer device, and storage medium for identifying persons having deformed spinal columns

Country Status (2)

Country Link
CN (1) CN111666890B (en)
WO (1) WO2021114623A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113505751A (en) * 2021-07-29 2021-10-15 同济大学 Human skeleton action recognition method based on difference map convolutional neural network
CN113610808A (en) * 2021-08-09 2021-11-05 中国科学院自动化研究所 Individual brain atlas individualization method, system and equipment based on individual brain connection atlas

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113139962B (en) * 2021-05-26 2021-11-30 北京欧应信息技术有限公司 System and method for scoliosis probability assessment
CN114287915B (en) * 2021-12-28 2024-03-05 深圳零动医疗科技有限公司 Noninvasive scoliosis screening method and system based on back color images

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3449822A1 (en) * 2016-04-25 2019-03-06 Telefield Medical Imaging Limited Method and device for measuring spinal column curvature
CN109431511A (en) * 2018-11-14 2019-03-08 南京航空航天大学 A kind of human body back scoliosis angle measurement method based on Digital Image Processing
CN109493334A (en) * 2018-11-12 2019-03-19 深圳码隆科技有限公司 Measure the method and device of spinal curvature
CN110415291A (en) * 2019-08-07 2019-11-05 清华大学 Image processing method and relevant device
CN110458831A (en) * 2019-08-12 2019-11-15 深圳市智影医疗科技有限公司 A kind of scoliosis image processing method based on deep learning

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108647588A (en) * 2018-04-24 2018-10-12 广州绿怡信息科技有限公司 Goods categories recognition methods, device, computer equipment and storage medium
CN109508638A (en) * 2018-10-11 2019-03-22 平安科技(深圳)有限公司 Face Emotion identification method, apparatus, computer equipment and storage medium
CN109657582B (en) * 2018-12-10 2023-10-31 平安科技(深圳)有限公司 Face emotion recognition method and device, computer equipment and storage medium
CN110781836A (en) * 2019-10-28 2020-02-11 深圳市赛为智能股份有限公司 Human body recognition method and device, computer equipment and storage medium
CN111144285B (en) * 2019-12-25 2024-06-14 中国平安人寿保险股份有限公司 Fat and thin degree identification method, device, equipment and medium
CN111191568B (en) * 2019-12-26 2024-06-14 中国平安人寿保险股份有限公司 Method, device, equipment and medium for identifying flip image

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3449822A1 (en) * 2016-04-25 2019-03-06 Telefield Medical Imaging Limited Method and device for measuring spinal column curvature
CN109493334A (en) * 2018-11-12 2019-03-19 深圳码隆科技有限公司 Measure the method and device of spinal curvature
CN109431511A (en) * 2018-11-14 2019-03-08 南京航空航天大学 A kind of human body back scoliosis angle measurement method based on Digital Image Processing
CN110415291A (en) * 2019-08-07 2019-11-05 清华大学 Image processing method and relevant device
CN110458831A (en) * 2019-08-12 2019-11-15 深圳市智影医疗科技有限公司 A kind of scoliosis image processing method based on deep learning

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113505751A (en) * 2021-07-29 2021-10-15 同济大学 Human skeleton action recognition method based on difference map convolutional neural network
CN113610808A (en) * 2021-08-09 2021-11-05 中国科学院自动化研究所 Individual brain atlas individualization method, system and equipment based on individual brain connection atlas
CN113610808B (en) * 2021-08-09 2023-11-03 中国科学院自动化研究所 Group brain map individuation method, system and equipment based on individual brain connection diagram

Also Published As

Publication number Publication date
CN111666890B (en) 2023-06-30
CN111666890A (en) 2020-09-15

Similar Documents

Publication Publication Date Title
WO2021114623A1 (en) Method, apparatus, computer device, and storage medium for identifying persons having deformed spinal columns
CN110120040B (en) Slice image processing method, slice image processing device, computer equipment and storage medium
Dua et al. Biometric iris recognition using radial basis function neural network
WO2020215557A1 (en) Medical image interpretation method and apparatus, computer device and storage medium
WO2020143309A1 (en) Segmentation model training method, oct image segmentation method and apparatus, device and medium
WO2020000908A1 (en) Method and device for face liveness detection
EP4075324A1 (en) Face recognition method and face recognition device
US11200665B2 (en) Fundus image processing method, computer apparatus, and storage medium
Flores et al. Segmentation of melanocytic skin lesions using feature learning and dictionaries
CN111583220B (en) Image data detection method and device
US11854200B2 (en) Skin abnormality monitoring systems and methods
Ichim et al. Melanoma detection using an objective system based on multiple connected neural networks
CN110838125A (en) Target detection method, device, equipment and storage medium of medical image
CN111368672A (en) Construction method and device for genetic disease facial recognition model
JP7294695B2 (en) Program, Information Recording Medium, Classification Apparatus, and Classification Method Based on Trained Model
WO2021120753A1 (en) Method and apparatus for recognition of luminal area in choroidal vessels, device, and medium
CN111178187A (en) Face recognition method and device based on convolutional neural network
CN115953824B (en) Face skin image processing method and system
CN113889267A (en) Method for constructing diabetes diagnosis model based on eye image recognition and electronic equipment
CN113781488A (en) Tongue picture image segmentation method, apparatus and medium
US20210073595A1 (en) Classification device, classification method, program, and information recording medium
CN113450305B (en) Medical image processing method, system, equipment and readable storage medium
Choudhary et al. Skin lesion detection based on deep neural networks
CN116563647A (en) Age-related maculopathy image classification method and device
US10956735B1 (en) System and method for determining a refractive error from red reflex images of eyes

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20900433

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20900433

Country of ref document: EP

Kind code of ref document: A1