CN114693682B - Spine feature identification method based on image processing - Google Patents
Spine feature identification method based on image processing Download PDFInfo
- Publication number
- CN114693682B CN114693682B CN202210610910.6A CN202210610910A CN114693682B CN 114693682 B CN114693682 B CN 114693682B CN 202210610910 A CN202210610910 A CN 202210610910A CN 114693682 B CN114693682 B CN 114693682B
- Authority
- CN
- China
- Prior art keywords
- frequency
- spine
- frequency pixel
- pixel points
- edge
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 28
- 238000012545 processing Methods 0.000 title claims abstract description 18
- 238000013507 mapping Methods 0.000 claims description 14
- 238000009825 accumulation Methods 0.000 claims description 9
- 238000001914 filtration Methods 0.000 claims description 7
- 230000001186 cumulative effect Effects 0.000 claims description 6
- 230000011218 segmentation Effects 0.000 claims description 6
- 238000005315 distribution function Methods 0.000 claims description 3
- 210000000988 bone and bone Anatomy 0.000 description 3
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 2
- 210000003205 muscle Anatomy 0.000 description 2
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 1
- XCWPUUGSGHNIDZ-UHFFFAOYSA-N Oxypertine Chemical compound C1=2C=C(OC)C(OC)=CC=2NC(C)=C1CCN(CC1)CCN1C1=CC=CC=C1 XCWPUUGSGHNIDZ-UHFFFAOYSA-N 0.000 description 1
- 208000020307 Spinal disease Diseases 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 229910052799 carbon Inorganic materials 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 201000010099 disease Diseases 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 238000004321 preservation Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/40—Image enhancement or restoration using histogram techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/94—Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/12—Edge-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20192—Edge enhancement; Edge preservation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30008—Bone
- G06T2207/30012—Spine; Backbone
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
- Facsimile Image Signal Circuits (AREA)
- Image Processing (AREA)
Abstract
The invention relates to the field of image processing, in particular to a spine feature identification method based on image processing.
Description
Technical Field
The application relates to the field of image processing, in particular to a spine feature identification method based on image processing.
Background
In the case of spinal diseases, medical images are required to be taken to diagnose the disease condition, and usually, due to the different sizes of medical equipment, most of the taken spinal images are line slices, and only the spinal conditions of local segments can be observed. The whole spine image can be shot by the standing type full spine image, but the larger the image is, the more complex the information is, the limited dynamic range of the photosensitive element is, the limited contrast of some key details is, and the image is not easy to observe manually.
Therefore, in order to observe a complete spine image more clearly, the invention optimizes and enhances the spine image characteristics in the spine medical image by using an image processing technology on the basis of machine vision to obtain an image with clear contrast of spine and bone textures and muscle contours, and image detail information is also kept, and the method is intelligent and accurate.
Disclosure of Invention
The invention provides a spine feature identification method based on image processing, which solves the problem of unclear full spine image in medical image and adopts the following technical scheme:
acquiring a full spine lateral image map, and performing semantic segmentation on the full spine lateral image map to obtain a full spine image map;
detecting the whole spine image map by using a sobel operator to obtain gradient edge pixel points of the whole spine image map;
carrying out frequency division filtering on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points of the whole spine image map;
determining spine edge high-frequency pixel points and non-spine edge high-frequency pixel points in the high-frequency pixel points by utilizing the gradient edge pixel points;
respectively taking the corresponding gray levels of the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points as the gray level of the spine edge high-frequency pixel points, the gray level of the non-spine edge high-frequency pixel points and the gray level of the low-frequency pixel points;
obtaining a gray level histogram according to the gray level of the spine edge high-frequency pixel point, the gray level of the non-spine edge high-frequency pixel point and the frequency of the gray level of the low-frequency pixel point in the whole spine image;
acquiring the average frequency of the gray levels of all low-frequency pixel points in the gray histogram and the frequency of the gray level of each high-frequency pixel point at the edge of the vertebra;
obtaining the weight ratio of the gray levels of all the low-frequency pixel points and the gray level of each spine edge high-frequency pixel point according to the average frequency of the gray levels of all the low-frequency pixel points and the frequency of the gray level of each spine edge high-frequency pixel point;
determining the gray levels of the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the weights of the gray levels of the non-spine edge high-frequency pixel points according to the gray levels of all the low-frequency pixel points and the weight ratio of the gray levels of each spine edge high-frequency pixel point;
and constructing an accumulation mapping function according to the gray levels of the low-frequency pixel points, the gray levels of the high-frequency pixel points at the spine edge and the gray levels of the high-frequency pixel points at the non-spine edge, and performing histogram equalization on the full spine image map by using the accumulation mapping function to obtain the processed full spine image map.
The method for determining the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points in the high-frequency pixel points comprises the following steps:
taking gradient edge pixel points of the full spine image map as spine edge high-frequency pixel points in the high-frequency pixel points;
and other pixels except the high-frequency pixels at the edge of the spine in the high-frequency pixels are non-high-frequency pixels at the edge of the spine.
The method for obtaining the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine comprises the following steps:
calculating the average frequency of the gray levels of all low-frequency pixel points in the gray level histogram;
obtaining the average value of the entropies of the gray levels of all the low-frequency pixel points according to the average frequency of the gray levels of all the low-frequency pixel points;
acquiring the frequency of the gray level of each high-frequency pixel point at the edge of each vertebra;
and taking the ratio of the frequency of the gray level of each spine edge high-frequency pixel point to the average value of the entropies of the gray levels of all low-frequency pixel points as the weight ratio of the gray level of each low-frequency pixel point to the gray level of each spine edge high-frequency pixel point.
The method for determining the weight of the gray level of the low-frequency pixel point, the gray level of the high-frequency pixel point at the edge of the spine and the gray level of the high-frequency pixel point at the edge of the non-spine comprises the following steps:
calculating a weight average value of the gray levels of the spine edge high-frequency pixel points according to the weight ratio of the gray levels of the low-frequency pixel points to the gray level of each spine edge high-frequency pixel point;
setting the ratio of the weight average value of the gray level of the spine edge high-frequency pixel points to the gray level weight of all non-spine edge high-frequency pixel points as 1: 2;
and calculating the gray level of the low-frequency pixel, the gray level of the spine edge high-frequency pixel and the weight of the gray level of the non-spine edge high-frequency pixel according to the gray level of the low-frequency pixel and the weight ratio of the gray level of each spine edge high-frequency pixel, the weight average value of the gray level of the spine edge high-frequency pixel and the gray level weight ratio of all non-spine edge high-frequency pixels.
The cumulative mapping function is as follows:
in the formula,the weight of the gray level of the high-frequency pixel point at the edge of the vertebra,the weight of the gray level of the non-vertebral edge high-frequency pixel point,is the weight of the gray level of the low-frequency pixel,for the gray scale cumulative distribution function of the histogram equalized image,is shown asThe mapping function of the original image and the equalized image at each gray level,is shown in the original imageThe frequency with which the individual gray levels occur,the total number of the pixel points is,for grey levels of the original image,As is the number of grey levels in the original image,for the purpose of the normalized gray scale level,,, ,respectively representing the gray levels of the original image as,,,The number of the pixel points.
The invention has the beneficial effects that: based on image processing, detecting the whole spine image map by using a sobel operator to obtain a gradient edge image of the whole spine image map, frequency division filtering is carried out on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points in the whole spine image map, dividing the high-frequency pixel points into spine edge high-frequency pixel points and non-spine edge high-frequency pixel points according to the high-frequency pixel points and the gradient edge pixel points of the whole spine image map, according to the frequency of the gray level distribution of the vertebra edge high-frequency pixel points, the non-vertebra edge high-frequency pixel points and the low-frequency pixel points in the gray level histogram, the gray levels of the high-frequency pixel points at the edge of the vertebra, the high-frequency pixel points at the edge of the non-vertebra and the low-frequency pixel points are subjected to weight distribution, the mapping accumulation function is obtained through weight distribution, histogram equalization is carried out on the whole spine image to obtain the whole spine image with clear contrast and complete details, and the method improves the identification degree of the spine medical image.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
Fig. 1 is a schematic flow chart of a spine feature identification method based on image processing according to the present invention;
fig. 2a is a schematic view of a full spine image of a spine feature recognition method based on image processing according to the present invention;
fig. 2b is a schematic diagram of a full spine image after histogram equalization in a spine feature identification method based on image processing according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
An embodiment of a spine feature identification method based on image processing according to the present invention is shown in fig. 1, and includes:
the method comprises the following steps: acquiring a lateral image map of the whole spine, and performing semantic segmentation to obtain the image map of the whole spine;
the purpose of this step is, gather the whole backbone image in the medical image, extract the backbone part image among them, as the basis of the subsequent data analysis.
It should be noted that: the whole spinal column film is obtained from a database of a hospital, comprises all bones from cervical vertebra to femoral shaft, and is formed by splicing a chest film and a waist film. But not a single chest film and a single waist film, because of multiple imaging, focal distance is difficult to align and coincide, vertebral body images of the two films cannot be connected, and angle measurement is inaccurate.
The semantic segmentation method comprises the following steps:
for the processing of the whole spine image map, the influence of non-target areas is removed as much as possible, and CNN semantic segmentation is adopted:
(1) CNN is Encoder-Decoder network, according to 7: a scale of 3 divides the data set into a training set and a test set.
(2) The spine region is marked as 1 and all other regions are marked with 0.
(3) The loss function used by the network is a cross entropy loss function.
The whole spine image and the background image can be obtained through semantic segmentation.
Step two: detecting the whole spine image map by using a sobel operator to obtain gradient edge pixel points of the whole spine image map; carrying out frequency division filtering on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points of the whole spine image map;
the purpose of this step is to carry out edge detection and frequency division on the image and classify the pixel points in the image.
It should be noted that, to improve the recognition effect of the spine image, the edge and texture details of the spine need to be enhanced, but a large amount of detail information of the spine is lost by directly performing histogram equalization processing on a complete full spine medical image, for example, fig. 2a is a full spine image, and fig. 2b is an image after histogram equalization of fig. 2a, it can be seen that a large amount of detail information is lost in the image, for example, contour information of excessively high brightness at the end of the lumbar vertebra is lost, the cervical vertebra is dark, the texture of the bony spur is lost, the bone edge becomes rough and unsmooth, and the like, because many pixels distributed less in the image are easily submerged, and because when the original image is mapped to the output image in the histogram equalization process, a small number of gray levels approximate to the accumulation result are rounded and combined, the gray scale range is compressed and therefore the detail features can be preserved as long as these low distributed gray levels are preserved.
In the embodiment, the detail information is processed by using a frequency division filter, the enhancement of the spine image is realized by preventing the gray levels of the parts from being combined and then equalizing, so that the contrast is enlarged and the details of key parts are kept as much as possible. The detailed characteristics of the spine can be well reserved so as to facilitate the recognition and extraction of the machine.
The method for acquiring the gradient edge pixel points of the whole spine image map comprises the following steps:
calculating each pixel point of the target area by using sobel operatorDirection andgradient of directionCounting all edge pixel point sets with gray gradient on vertebra。
The method for acquiring the high-frequency pixel points and the low-frequency pixel points in the whole spine image map comprises the following steps:
use the branchImage filteringDivided into low-frequency partsAnd high frequencyBecause a large amount of noise is generated in the process of medical image shooting, the types of the noise are various, some noises can blur images, some noises can gather to become dead points to destroy details of the images, and the noises in the images can be amplified during histogram equalization, the embodiment mainly protects the edge contour and detail information of the images, so that the problem of blurring the details by other linear filtering can be solved by adopting a median filtering algorithm.
Step three: determining spine edge high-frequency pixel points and non-spine edge high-frequency pixel points in the high-frequency pixel points by utilizing the gradient edge pixel points;
the purpose of the step is to classify the pixel points in the high-frequency information of the image, store the detail characteristics in the image,
the method for dividing the high-frequency pixel points into the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points comprises the following steps:
gradient edge pixel point set using a full spine image mapTo eliminate high frequency partOf non-spinal edge partsNamely:
wherein,is a high-frequency pixel point at the edge of the vertebra,is a high-frequency pixel point at the edge of non-vertebra,are high frequency pixels.
Step four: taking the gray levels corresponding to the vertebra edge high-frequency pixel points, the non-vertebra edge high-frequency pixel points and the low-frequency pixel points as the gray levels of the vertebra edge high-frequency pixel points, the non-vertebra edge high-frequency pixel points and the low-frequency pixel points respectively; obtaining a gray level histogram according to the gray level of the vertebra edge high-frequency pixel points, the gray level of the non-vertebra edge high-frequency pixel points and the frequency of the gray level of the low-frequency pixel points appearing in the whole vertebra image map; acquiring the average frequency of the gray levels of all low-frequency pixel points in the gray histogram and the frequency of the gray level of each high-frequency pixel point at the edge of the vertebra; obtaining the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each spine edge high-frequency pixel point according to the average frequency of the gray levels of all the low-frequency pixel points and the frequency of the gray level of each spine edge high-frequency pixel point; determining the gray levels of the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the weights of the gray levels of the non-spine edge high-frequency pixel points according to the gray levels of all the low-frequency pixel points and the weight ratio of the gray levels of each spine edge high-frequency pixel point;
the purpose of this step is to calculate the weight of the grey level according to the frequency of the grey level corresponding to the different kinds of pixel points in the histogram.
The gray levels of the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points are gray levels corresponding to the gray values of the spine edge high-frequency pixel points, the gray levels corresponding to the gray values of the non-spine edge high-frequency pixel points and the gray levels corresponding to the gray values of the low-frequency pixel points, and the gray levels of the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points are determined according to the gray levels of the spine edge high-frequency pixel points and the gray levels corresponding to the gray values of the low-frequency pixel pointsObtaining the gray histogram of the gray levels of the high-frequency pixel points at the vertebra edge and the frequency of the gray levels of the low-frequency pixel points in the full vertebra image map, and giving weight to the gray levels of the high-frequency pixel points at the vertebra edgeWeighting the gray level of the high-frequency pixel point at the edge of non-vertebraWeighting the gray level of the low-frequency pixel。
The method for acquiring the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine comprises the following steps:
(1) obtaining the frequency of each high-frequency vertebra gray level appearing in the gray histogramAnd the average frequency of the gray levels of all low-frequency pixel points appearing in the gray histogram;
(3) Obtaining the average value of the entropies of the gray levels of all the low-frequency pixel points according to the average frequency of the gray levels of all the low-frequency pixel points appearing in the gray histogram;
(3) Entropy average value according to gray level of low-frequency pixel pointAnd each of the high-frequency pixel gray levels at the edge of the spineObtaining the gray level of each high-frequency pixel point at the edge of each vertebra by the ratio of the gray levelsAnd low frequency pixel gray scaleThe weight ratio of (A):
in the formula, the content of the active carbon is shown in the specification,for each vertebra edge high frequency pixel gray levelAnd low frequency pixel gray scaleThe weight ratio of (a) to (b),entropy average of gray levels of low-frequency pixel pointsThe ratio of the gray level of the high-frequency pixel point at the edge of each vertebra,for the frequency of occurrence of the grey levels,the frequency of the gray level of the high-frequency pixel points at the edge of the spine,the average frequency of the gray levels of the low-frequency pixel points is shown, and when the occurrence frequency of the gray levels of the high-frequency pixel points at the edge of the spine is equal to the average frequency of the gray levels of the low frequency, the gray levels can be kept as much as most of low-frequency information.
It should be noted that, first, for the high-frequency edge portion with the highest importance degree, the histogram equalization is the most ideal state when the frequency of occurrence of each gray level is equal. However, the total number of the pixels is fixed, and many low-distribution gray levels will inevitably disappear due to combination during histogram equalization, so that the gray levels combined in the low-frequency information do not need to be considered, and only the high-frequency gray levels are subjected toAndthe remaining processing is performed, only the frequency is adjusted, which inevitably causes the uncertainty of gray level combination, if the average frequency of the gray levels of the low-frequency pixels is too large, in order to make the average frequency larger=More low-profile gray levels are merged, which ignores the richness of the gray levels. The entropy is used for describing the disorder degree of the system, the larger the entropy is, the larger the uncertainty of the system is, so that the larger the image information amount is, and after each high-frequency low-distribution gray level which is easy to swallow is given a weight, the entropy value of each high-frequency low-distribution gray level is equal to the average value of the gray level entropy of the low-frequency pixel point, so that the high-frequency gray levels can be reserved.
The method for obtaining the gray levels of the low-frequency pixel points, the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points comprises the following steps:
(1) this embodiment is directed to non-spinal edge high frequency pixelCorresponding non-spine edge high frequency pixel gray scaleEndowing a uniform weight valueCalculating the weight mean value of the gray levels of all the vertebra edge high-frequency pixel points, and weighting the gray levels of the non-vertebra edge high-frequency pixel pointsAnd the weight mean value of the gray levels of all the high-frequency pixel points at the edge of the spineIs set to;
Because forIt contains not only high-frequency information on the spine but also high-frequency information on muscle texture, organs, and the like. And therefore less important than the spine edge, a portion of the gray levels may be "sacrificed" to ensure preservation of the spine edge gray levels during histogram equalization.
(2) According to the weight ratio of the gray levels of all low-frequency pixel points and the gray level of each high-frequency pixel point at the edge of the vertebra, namelyAndthe ratio, the proportional relation between the gray level weight of the non-spinal edge high-frequency pixel and the weight mean value of the gray levels of all spinal edge high-frequency pixelsAndthe ratio of the gray scale weight of the high-frequency pixel at the spine edge to the gray scale weight of the low-frequency pixel and the gray scale of the high-frequency pixel at the non-spine edge can be obtained、、And carrying out weight distribution according to the proportional relation.
For example, the calculation method of the gray level weight of the spine edge high-frequency pixel point, the gray level weight of the low-frequency pixel point and the gray level weight of the non-spine edge high-frequency pixel point is as follows:
(1) suppose the gray level of the high-frequency pixel point at the edge of the vertebra is、The gray level of the non-vertebral edge high-frequency pixel point is、Low frequency pixel point gray scale of;
(2) Obtaining the weight ratio of the gray levels of all the low-frequency pixel points and the gray level of each high-frequency pixel point at the edge of the vertebra according to the stepThe weight ratio of the gray level of the point to the gray level of the high-frequency pixel point at the edge of each vertebra is, Then, then=, =;
(3) Obtaining the proportion of the gray level weight of the non-spine edge high-frequency pixel points to the weight average of the gray levels of all spine edge high-frequency pixel points according to the step:;wherein=Then, then,;
(4) Then obtainObtaining the gray level of the high-frequency pixel point at the edge of the vertebra according to the proportional relation、High frequency gray scale of non-vertebral edge pixel、Gray level of low frequency pixelThe weight of (c):
Step five: and constructing an accumulation mapping function according to the gray level of the low-frequency pixel point, the gray level of the spine edge high-frequency pixel point and the weight of the gray level of the non-spine edge high-frequency pixel point, and performing histogram equalization on the whole spine image map by using the accumulation mapping function to obtain the processed whole spine image map.
The purpose of the step is to modify the mapping accumulation function of histogram equalization according to the gray level weight of the non-vertebra edge high-frequency pixel point, the gray level weight of the vertebra edge high-frequency pixel point and the gray level weight of the low-frequency pixel point, perform histogram equalization and enhance the image.
The cumulative mapping function after adding weight distribution during histogram equalization is as follows:
in the formula, the first step is that,the weight of the gray level of the high-frequency pixel point at the edge of the vertebra,the weight of the gray level of the non-vertebral edge high-frequency pixel point,is the weight of the gray level of the low-frequency pixel,for the gray scale cumulative distribution function of the histogram equalized image,denotes the firstThe mapping function of the original image and the equalized image at each gray level,is shown in the original imageThe frequency with which the individual gray levels occur,the total number of the pixel points is,for grey levels of the original image,As is the number of grey levels in the original image,for the purpose of the normalized gray scale level,,, ,respectively representing the gray levels of the original image as,,,The number of the pixel points.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.
Claims (2)
1. A spine feature identification method based on image processing is characterized by comprising the following steps:
acquiring a full spine lateral image map, and performing semantic segmentation on the full spine lateral image map to obtain a full spine image map;
detecting the whole spine image map by using a sobel operator to obtain gradient edge pixel points of the whole spine image map;
carrying out frequency division filtering on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points of the whole spine image map;
determining spine edge high-frequency pixel points and non-spine edge high-frequency pixel points in the high-frequency pixel points by utilizing the gradient edge pixel points;
taking the gray levels corresponding to the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points as the gray levels of the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points respectively;
obtaining a gray level histogram according to the gray level of the vertebra edge high-frequency pixel points, the gray level of the non-vertebra edge high-frequency pixel points and the frequency of the gray level of the low-frequency pixel points appearing in the whole vertebra image map;
acquiring the average frequency of the gray levels of all low-frequency pixel points in the gray histogram and the frequency of the gray level of each high-frequency pixel point at the edge of the vertebra;
obtaining the weight ratio of the gray levels of all the low-frequency pixel points and the gray level of each spine edge high-frequency pixel point according to the average frequency of the gray levels of all the low-frequency pixel points and the frequency of the gray level of each spine edge high-frequency pixel point;
the method for obtaining the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine comprises the following steps:
calculating the average frequency of the gray levels of all low-frequency pixel points in the gray histogram;
obtaining the average value of the entropies of the gray levels of all the low-frequency pixel points according to the average frequency of the gray levels of all the low-frequency pixel points;
acquiring the frequency of the gray level of each high-frequency pixel point at the edge of each vertebra;
taking the ratio of the frequency of the gray level of each spine edge high-frequency pixel point to the average value of the entropies of the gray levels of all low-frequency pixel points as the weight ratio of the gray level of each spine edge high-frequency pixel point to the gray level of each low-frequency pixel point;
determining the gray levels of all the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the weights of the gray levels of the non-spine edge high-frequency pixel points according to the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each spine edge high-frequency pixel point;
the method for determining the weight of the gray level of the low-frequency pixel point, the gray level of the high-frequency pixel point at the edge of the spine and the gray level of the high-frequency pixel point at the edge of the non-spine comprises the following steps:
calculating a weight average value of the gray levels of the high-frequency pixel points at the edge of the spine according to the weight ratio of the gray levels of the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine;
setting the ratio of the weight mean value of the gray level of the high-frequency pixel points at the edge of the vertebra to the gray level weight of all the high-frequency pixel points at the edge of the non-vertebra as;
Obtaining the proportional relation of the gray level of the low-frequency pixel point, the gray level of the spine edge high-frequency pixel point and the gray level of the non-spine edge high-frequency pixel point according to the weight ratio of the gray level of the low-frequency pixel point to the gray level of each spine edge high-frequency pixel point, the weight average value of the gray level of the spine edge high-frequency pixel point and the weight proportion of the gray levels of all non-spine edge high-frequency pixel points;
carrying out weight distribution according to the proportional relation of the gray level of the low-frequency pixel point, the gray level of the spine edge high-frequency pixel point and the gray level of the non-spine edge high-frequency pixel point;
constructing an accumulation mapping function according to the gray levels of the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the gray levels of the non-spine edge high-frequency pixel points, and performing histogram equalization on the whole spine image map by using the accumulation mapping function to obtain a processed whole spine image map;
the cumulative mapping function is as follows:
in the formula,the weight of the gray level of the high-frequency pixel point at the edge of the vertebra,the weight of the gray level of the non-vertebral edge high-frequency pixel point,is the weight of the gray level of the low-frequency pixel,for the gray scale cumulative distribution function of the histogram equalized image,denotes the firstThe mapping function of the original image and the equalized image at each gray level,is shown in the original imageThe frequency with which the individual gray levels occur,the total number of the pixel points is,for grey levels of the original image,As is the number of grey levels in the original image,for the purpose of the normalized gray scale level,,, ,respectively representing the gray levels of the original image as,,,The number of the pixel points.
2. The spine feature identification method based on image processing according to claim 1, wherein the method for determining spine edge high frequency pixel points and non-spine edge high frequency pixel points among the high frequency pixel points comprises:
taking gradient edge pixel points of the full spine image map as spine edge high-frequency pixel points in the high-frequency pixel points;
and other pixels except the high-frequency pixels at the edge of the spine in the high-frequency pixels are non-high-frequency pixels at the edge of the spine.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210610910.6A CN114693682B (en) | 2022-06-01 | 2022-06-01 | Spine feature identification method based on image processing |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210610910.6A CN114693682B (en) | 2022-06-01 | 2022-06-01 | Spine feature identification method based on image processing |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114693682A CN114693682A (en) | 2022-07-01 |
CN114693682B true CN114693682B (en) | 2022-08-26 |
Family
ID=82131266
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210610910.6A Active CN114693682B (en) | 2022-06-01 | 2022-06-01 | Spine feature identification method based on image processing |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114693682B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115830459B (en) * | 2023-02-14 | 2023-05-12 | 山东省国土空间生态修复中心(山东省地质灾害防治技术指导中心、山东省土地储备中心) | Mountain forest grass life community damage degree detection method based on neural network |
CN117745722B (en) * | 2024-02-20 | 2024-04-30 | 北京大学 | Medical health physical examination big data optimization enhancement method |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10336657A (en) * | 1997-05-29 | 1998-12-18 | Ricoh Co Ltd | Image processor |
JPH11136521A (en) * | 1997-10-31 | 1999-05-21 | Ricoh Co Ltd | Picture data processor |
CN106651818A (en) * | 2016-11-07 | 2017-05-10 | 湖南源信光电科技有限公司 | Improved Histogram equalization low-illumination image enhancement algorithm |
CN108230260A (en) * | 2017-12-06 | 2018-06-29 | 天津津航计算技术研究所 | A kind of fusion method of new infrared image and twilight image |
CN109919929A (en) * | 2019-03-06 | 2019-06-21 | 电子科技大学 | A kind of fissuring of tongue feature extracting method based on wavelet transformation |
WO2020103601A1 (en) * | 2018-11-21 | 2020-05-28 | Zhejiang Dahua Technology Co., Ltd. | Method and system for generating a fusion image |
CN111899205A (en) * | 2020-08-10 | 2020-11-06 | 国科天成(北京)科技有限公司 | Image enhancement method of scene self-adaptive wide dynamic infrared thermal imaging |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110852977B (en) * | 2019-10-29 | 2023-04-11 | 天津大学 | Image enhancement method for fusing edge gray level histogram and human eye visual perception characteristics |
CN114494256B (en) * | 2022-04-14 | 2022-06-14 | 武汉金龙电线电缆有限公司 | Electric wire production defect detection method based on image processing |
-
2022
- 2022-06-01 CN CN202210610910.6A patent/CN114693682B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10336657A (en) * | 1997-05-29 | 1998-12-18 | Ricoh Co Ltd | Image processor |
JPH11136521A (en) * | 1997-10-31 | 1999-05-21 | Ricoh Co Ltd | Picture data processor |
CN106651818A (en) * | 2016-11-07 | 2017-05-10 | 湖南源信光电科技有限公司 | Improved Histogram equalization low-illumination image enhancement algorithm |
CN108230260A (en) * | 2017-12-06 | 2018-06-29 | 天津津航计算技术研究所 | A kind of fusion method of new infrared image and twilight image |
WO2020103601A1 (en) * | 2018-11-21 | 2020-05-28 | Zhejiang Dahua Technology Co., Ltd. | Method and system for generating a fusion image |
CN109919929A (en) * | 2019-03-06 | 2019-06-21 | 电子科技大学 | A kind of fissuring of tongue feature extracting method based on wavelet transformation |
CN111899205A (en) * | 2020-08-10 | 2020-11-06 | 国科天成(北京)科技有限公司 | Image enhancement method of scene self-adaptive wide dynamic infrared thermal imaging |
Non-Patent Citations (3)
Title |
---|
PET/CT肺部成像过程中的衰减校正和分割方法的研究;张倩,;《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》;20140215;第2014年卷(第2期);第I138-808页 * |
Research on feature extraction algorithm of pavement disease;Lixia Cao 等,;《2021 International Conference on Electronic Information Engineering and Computer Science (EIECS)》;20211113;第2021卷;第1-4页 * |
耦合边缘检测与优化的多尺度遥感图像融合法;谷志鹏 等,;《计算机工程与应用》;20170601;第53卷(第11期);第192-198页 * |
Also Published As
Publication number | Publication date |
---|---|
CN114693682A (en) | 2022-07-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114648530B (en) | CT image processing method | |
CN114693682B (en) | Spine feature identification method based on image processing | |
WO2021017297A1 (en) | Artificial intelligence-based spine image processing method and related device | |
US7218763B2 (en) | Method for automated window-level settings for magnetic resonance images | |
CN103026379B (en) | The method calculating image noise level | |
CN112446880B (en) | Image processing method, electronic device and readable storage medium | |
CN109410177A (en) | A kind of image quality analysis method and system of super-resolution image | |
CN117237591A (en) | Intelligent removal method for heart ultrasonic image artifacts | |
CN117876402B (en) | Intelligent segmentation method for temporomandibular joint disorder image | |
CN111951215A (en) | Image detection method and device and computer readable storage medium | |
CN104574392B (en) | A kind of retrogression computer automatic grading method of interverbebral disc image | |
CN108038840B (en) | Image processing method and device, image processing equipment and storage medium | |
CN117408988A (en) | Artificial intelligence-based focus image analysis method and apparatus | |
CN118333893A (en) | Preoperative blood vessel assessment method for free flap surgery | |
CN118229538A (en) | Intelligent enhancement method for bone quality CT image | |
CN117237342B (en) | Intelligent analysis method for respiratory rehabilitation CT image | |
CN114332255B (en) | Medical image processing method and device | |
CN116029934A (en) | Low-dose DR image and CT image denoising method | |
Yang et al. | Fusion of CT and MR images using an improved wavelet based method | |
Kumar et al. | Semiautomatic method for segmenting pedicles in vertebral radiographs | |
CN115205241A (en) | Metering method and system for apparent cell density | |
CN110084770B (en) | Brain image fusion method based on two-dimensional Littlewood-Paley empirical wavelet transform | |
Pancholi et al. | A Review of Noise Reduction Filtering Techniques for MRI Images | |
CN114418920B (en) | Endoscope multi-focus image fusion method | |
CN118570205B (en) | Image processing method, device and system based on portable X-ray imaging |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |