CN114693682B - Spine feature identification method based on image processing - Google Patents

Spine feature identification method based on image processing Download PDF

Info

Publication number
CN114693682B
CN114693682B CN202210610910.6A CN202210610910A CN114693682B CN 114693682 B CN114693682 B CN 114693682B CN 202210610910 A CN202210610910 A CN 202210610910A CN 114693682 B CN114693682 B CN 114693682B
Authority
CN
China
Prior art keywords
frequency
spine
frequency pixel
pixel points
edge
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210610910.6A
Other languages
Chinese (zh)
Other versions
CN114693682A (en
Inventor
马学晓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Affiliated Hospital of University of Qingdao
Original Assignee
Affiliated Hospital of University of Qingdao
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Affiliated Hospital of University of Qingdao filed Critical Affiliated Hospital of University of Qingdao
Priority to CN202210610910.6A priority Critical patent/CN114693682B/en
Publication of CN114693682A publication Critical patent/CN114693682A/en
Application granted granted Critical
Publication of CN114693682B publication Critical patent/CN114693682B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • G06T5/94Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20192Edge enhancement; Edge preservation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30008Bone
    • G06T2207/30012Spine; Backbone

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)
  • Facsimile Image Signal Circuits (AREA)
  • Image Processing (AREA)

Abstract

The invention relates to the field of image processing, in particular to a spine feature identification method based on image processing.

Description

Spine feature identification method based on image processing
Technical Field
The application relates to the field of image processing, in particular to a spine feature identification method based on image processing.
Background
In the case of spinal diseases, medical images are required to be taken to diagnose the disease condition, and usually, due to the different sizes of medical equipment, most of the taken spinal images are line slices, and only the spinal conditions of local segments can be observed. The whole spine image can be shot by the standing type full spine image, but the larger the image is, the more complex the information is, the limited dynamic range of the photosensitive element is, the limited contrast of some key details is, and the image is not easy to observe manually.
Therefore, in order to observe a complete spine image more clearly, the invention optimizes and enhances the spine image characteristics in the spine medical image by using an image processing technology on the basis of machine vision to obtain an image with clear contrast of spine and bone textures and muscle contours, and image detail information is also kept, and the method is intelligent and accurate.
Disclosure of Invention
The invention provides a spine feature identification method based on image processing, which solves the problem of unclear full spine image in medical image and adopts the following technical scheme:
acquiring a full spine lateral image map, and performing semantic segmentation on the full spine lateral image map to obtain a full spine image map;
detecting the whole spine image map by using a sobel operator to obtain gradient edge pixel points of the whole spine image map;
carrying out frequency division filtering on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points of the whole spine image map;
determining spine edge high-frequency pixel points and non-spine edge high-frequency pixel points in the high-frequency pixel points by utilizing the gradient edge pixel points;
respectively taking the corresponding gray levels of the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points as the gray level of the spine edge high-frequency pixel points, the gray level of the non-spine edge high-frequency pixel points and the gray level of the low-frequency pixel points;
obtaining a gray level histogram according to the gray level of the spine edge high-frequency pixel point, the gray level of the non-spine edge high-frequency pixel point and the frequency of the gray level of the low-frequency pixel point in the whole spine image;
acquiring the average frequency of the gray levels of all low-frequency pixel points in the gray histogram and the frequency of the gray level of each high-frequency pixel point at the edge of the vertebra;
obtaining the weight ratio of the gray levels of all the low-frequency pixel points and the gray level of each spine edge high-frequency pixel point according to the average frequency of the gray levels of all the low-frequency pixel points and the frequency of the gray level of each spine edge high-frequency pixel point;
determining the gray levels of the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the weights of the gray levels of the non-spine edge high-frequency pixel points according to the gray levels of all the low-frequency pixel points and the weight ratio of the gray levels of each spine edge high-frequency pixel point;
and constructing an accumulation mapping function according to the gray levels of the low-frequency pixel points, the gray levels of the high-frequency pixel points at the spine edge and the gray levels of the high-frequency pixel points at the non-spine edge, and performing histogram equalization on the full spine image map by using the accumulation mapping function to obtain the processed full spine image map.
The method for determining the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points in the high-frequency pixel points comprises the following steps:
taking gradient edge pixel points of the full spine image map as spine edge high-frequency pixel points in the high-frequency pixel points;
and other pixels except the high-frequency pixels at the edge of the spine in the high-frequency pixels are non-high-frequency pixels at the edge of the spine.
The method for obtaining the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine comprises the following steps:
calculating the average frequency of the gray levels of all low-frequency pixel points in the gray level histogram;
obtaining the average value of the entropies of the gray levels of all the low-frequency pixel points according to the average frequency of the gray levels of all the low-frequency pixel points;
acquiring the frequency of the gray level of each high-frequency pixel point at the edge of each vertebra;
and taking the ratio of the frequency of the gray level of each spine edge high-frequency pixel point to the average value of the entropies of the gray levels of all low-frequency pixel points as the weight ratio of the gray level of each low-frequency pixel point to the gray level of each spine edge high-frequency pixel point.
The method for determining the weight of the gray level of the low-frequency pixel point, the gray level of the high-frequency pixel point at the edge of the spine and the gray level of the high-frequency pixel point at the edge of the non-spine comprises the following steps:
calculating a weight average value of the gray levels of the spine edge high-frequency pixel points according to the weight ratio of the gray levels of the low-frequency pixel points to the gray level of each spine edge high-frequency pixel point;
setting the ratio of the weight average value of the gray level of the spine edge high-frequency pixel points to the gray level weight of all non-spine edge high-frequency pixel points as 1: 2;
and calculating the gray level of the low-frequency pixel, the gray level of the spine edge high-frequency pixel and the weight of the gray level of the non-spine edge high-frequency pixel according to the gray level of the low-frequency pixel and the weight ratio of the gray level of each spine edge high-frequency pixel, the weight average value of the gray level of the spine edge high-frequency pixel and the gray level weight ratio of all non-spine edge high-frequency pixels.
The cumulative mapping function is as follows:
Figure 811731DEST_PATH_IMAGE001
in the formula,
Figure 772734DEST_PATH_IMAGE002
the weight of the gray level of the high-frequency pixel point at the edge of the vertebra,
Figure 545518DEST_PATH_IMAGE003
the weight of the gray level of the non-vertebral edge high-frequency pixel point,
Figure 867913DEST_PATH_IMAGE004
is the weight of the gray level of the low-frequency pixel,
Figure 42543DEST_PATH_IMAGE005
for the gray scale cumulative distribution function of the histogram equalized image,
Figure 908867DEST_PATH_IMAGE006
is shown as
Figure 168948DEST_PATH_IMAGE007
The mapping function of the original image and the equalized image at each gray level,
Figure 477391DEST_PATH_IMAGE008
is shown in the original image
Figure 772107DEST_PATH_IMAGE009
The frequency with which the individual gray levels occur,
Figure 74912DEST_PATH_IMAGE010
the total number of the pixel points is,
Figure 822288DEST_PATH_IMAGE007
for grey levels of the original image
Figure 235952DEST_PATH_IMAGE011
Figure 621059DEST_PATH_IMAGE012
As is the number of grey levels in the original image,
Figure 94766DEST_PATH_IMAGE013
for the purpose of the normalized gray scale level,
Figure 63859DEST_PATH_IMAGE014
Figure 15634DEST_PATH_IMAGE015
,
Figure 340728DEST_PATH_IMAGE016
Figure 482528DEST_PATH_IMAGE017
respectively representing the gray levels of the original image as
Figure 673338DEST_PATH_IMAGE018
Figure 428804DEST_PATH_IMAGE019
Figure 522925DEST_PATH_IMAGE020
Figure 338434DEST_PATH_IMAGE007
The number of the pixel points.
The invention has the beneficial effects that: based on image processing, detecting the whole spine image map by using a sobel operator to obtain a gradient edge image of the whole spine image map, frequency division filtering is carried out on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points in the whole spine image map, dividing the high-frequency pixel points into spine edge high-frequency pixel points and non-spine edge high-frequency pixel points according to the high-frequency pixel points and the gradient edge pixel points of the whole spine image map, according to the frequency of the gray level distribution of the vertebra edge high-frequency pixel points, the non-vertebra edge high-frequency pixel points and the low-frequency pixel points in the gray level histogram, the gray levels of the high-frequency pixel points at the edge of the vertebra, the high-frequency pixel points at the edge of the non-vertebra and the low-frequency pixel points are subjected to weight distribution, the mapping accumulation function is obtained through weight distribution, histogram equalization is carried out on the whole spine image to obtain the whole spine image with clear contrast and complete details, and the method improves the identification degree of the spine medical image.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
Fig. 1 is a schematic flow chart of a spine feature identification method based on image processing according to the present invention;
fig. 2a is a schematic view of a full spine image of a spine feature recognition method based on image processing according to the present invention;
fig. 2b is a schematic diagram of a full spine image after histogram equalization in a spine feature identification method based on image processing according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
An embodiment of a spine feature identification method based on image processing according to the present invention is shown in fig. 1, and includes:
the method comprises the following steps: acquiring a lateral image map of the whole spine, and performing semantic segmentation to obtain the image map of the whole spine;
the purpose of this step is, gather the whole backbone image in the medical image, extract the backbone part image among them, as the basis of the subsequent data analysis.
It should be noted that: the whole spinal column film is obtained from a database of a hospital, comprises all bones from cervical vertebra to femoral shaft, and is formed by splicing a chest film and a waist film. But not a single chest film and a single waist film, because of multiple imaging, focal distance is difficult to align and coincide, vertebral body images of the two films cannot be connected, and angle measurement is inaccurate.
The semantic segmentation method comprises the following steps:
for the processing of the whole spine image map, the influence of non-target areas is removed as much as possible, and CNN semantic segmentation is adopted:
(1) CNN is Encoder-Decoder network, according to 7: a scale of 3 divides the data set into a training set and a test set.
(2) The spine region is marked as 1 and all other regions are marked with 0.
(3) The loss function used by the network is a cross entropy loss function.
The whole spine image and the background image can be obtained through semantic segmentation.
Step two: detecting the whole spine image map by using a sobel operator to obtain gradient edge pixel points of the whole spine image map; carrying out frequency division filtering on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points of the whole spine image map;
the purpose of this step is to carry out edge detection and frequency division on the image and classify the pixel points in the image.
It should be noted that, to improve the recognition effect of the spine image, the edge and texture details of the spine need to be enhanced, but a large amount of detail information of the spine is lost by directly performing histogram equalization processing on a complete full spine medical image, for example, fig. 2a is a full spine image, and fig. 2b is an image after histogram equalization of fig. 2a, it can be seen that a large amount of detail information is lost in the image, for example, contour information of excessively high brightness at the end of the lumbar vertebra is lost, the cervical vertebra is dark, the texture of the bony spur is lost, the bone edge becomes rough and unsmooth, and the like, because many pixels distributed less in the image are easily submerged, and because when the original image is mapped to the output image in the histogram equalization process, a small number of gray levels approximate to the accumulation result are rounded and combined, the gray scale range is compressed and therefore the detail features can be preserved as long as these low distributed gray levels are preserved.
In the embodiment, the detail information is processed by using a frequency division filter, the enhancement of the spine image is realized by preventing the gray levels of the parts from being combined and then equalizing, so that the contrast is enlarged and the details of key parts are kept as much as possible. The detailed characteristics of the spine can be well reserved so as to facilitate the recognition and extraction of the machine.
The method for acquiring the gradient edge pixel points of the whole spine image map comprises the following steps:
calculating each pixel point of the target area by using sobel operator
Figure 282119DEST_PATH_IMAGE021
Direction and
Figure 805327DEST_PATH_IMAGE022
gradient of direction
Figure 314805DEST_PATH_IMAGE023
Counting all edge pixel point sets with gray gradient on vertebra
Figure 35637DEST_PATH_IMAGE024
The method for acquiring the high-frequency pixel points and the low-frequency pixel points in the whole spine image map comprises the following steps:
use the branchImage filtering
Figure 466618DEST_PATH_IMAGE025
Divided into low-frequency parts
Figure 563887DEST_PATH_IMAGE026
And high frequency
Figure 599976DEST_PATH_IMAGE027
Because a large amount of noise is generated in the process of medical image shooting, the types of the noise are various, some noises can blur images, some noises can gather to become dead points to destroy details of the images, and the noises in the images can be amplified during histogram equalization, the embodiment mainly protects the edge contour and detail information of the images, so that the problem of blurring the details by other linear filtering can be solved by adopting a median filtering algorithm.
Step three: determining spine edge high-frequency pixel points and non-spine edge high-frequency pixel points in the high-frequency pixel points by utilizing the gradient edge pixel points;
the purpose of the step is to classify the pixel points in the high-frequency information of the image, store the detail characteristics in the image,
the method for dividing the high-frequency pixel points into the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points comprises the following steps:
gradient edge pixel point set using a full spine image map
Figure 491709DEST_PATH_IMAGE028
To eliminate high frequency part
Figure 908521DEST_PATH_IMAGE027
Of non-spinal edge parts
Figure 809481DEST_PATH_IMAGE029
Namely:
Figure 285084DEST_PATH_IMAGE030
wherein,
Figure 347718DEST_PATH_IMAGE031
is a high-frequency pixel point at the edge of the vertebra,
Figure 487712DEST_PATH_IMAGE029
is a high-frequency pixel point at the edge of non-vertebra,
Figure 192363DEST_PATH_IMAGE027
are high frequency pixels.
Step four: taking the gray levels corresponding to the vertebra edge high-frequency pixel points, the non-vertebra edge high-frequency pixel points and the low-frequency pixel points as the gray levels of the vertebra edge high-frequency pixel points, the non-vertebra edge high-frequency pixel points and the low-frequency pixel points respectively; obtaining a gray level histogram according to the gray level of the vertebra edge high-frequency pixel points, the gray level of the non-vertebra edge high-frequency pixel points and the frequency of the gray level of the low-frequency pixel points appearing in the whole vertebra image map; acquiring the average frequency of the gray levels of all low-frequency pixel points in the gray histogram and the frequency of the gray level of each high-frequency pixel point at the edge of the vertebra; obtaining the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each spine edge high-frequency pixel point according to the average frequency of the gray levels of all the low-frequency pixel points and the frequency of the gray level of each spine edge high-frequency pixel point; determining the gray levels of the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the weights of the gray levels of the non-spine edge high-frequency pixel points according to the gray levels of all the low-frequency pixel points and the weight ratio of the gray levels of each spine edge high-frequency pixel point;
the purpose of this step is to calculate the weight of the grey level according to the frequency of the grey level corresponding to the different kinds of pixel points in the histogram.
The gray levels of the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points are gray levels corresponding to the gray values of the spine edge high-frequency pixel points, the gray levels corresponding to the gray values of the non-spine edge high-frequency pixel points and the gray levels corresponding to the gray values of the low-frequency pixel points, and the gray levels of the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points are determined according to the gray levels of the spine edge high-frequency pixel points and the gray levels corresponding to the gray values of the low-frequency pixel pointsObtaining the gray histogram of the gray levels of the high-frequency pixel points at the vertebra edge and the frequency of the gray levels of the low-frequency pixel points in the full vertebra image map, and giving weight to the gray levels of the high-frequency pixel points at the vertebra edge
Figure 999782DEST_PATH_IMAGE032
Weighting the gray level of the high-frequency pixel point at the edge of non-vertebra
Figure 498896DEST_PATH_IMAGE033
Weighting the gray level of the low-frequency pixel
Figure 126187DEST_PATH_IMAGE034
The method for acquiring the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine comprises the following steps:
(1) obtaining the frequency of each high-frequency vertebra gray level appearing in the gray histogram
Figure 368949DEST_PATH_IMAGE035
And the average frequency of the gray levels of all low-frequency pixel points appearing in the gray histogram
Figure 266760DEST_PATH_IMAGE036
(3) Obtaining the average value of the entropies of the gray levels of all the low-frequency pixel points according to the average frequency of the gray levels of all the low-frequency pixel points appearing in the gray histogram
Figure 140038DEST_PATH_IMAGE037
Figure 717210DEST_PATH_IMAGE038
(3) Entropy average value according to gray level of low-frequency pixel point
Figure 999549DEST_PATH_IMAGE037
And each of the high-frequency pixel gray levels at the edge of the spineObtaining the gray level of each high-frequency pixel point at the edge of each vertebra by the ratio of the gray levels
Figure DEST_PATH_IMAGE039
And low frequency pixel gray scale
Figure 47140DEST_PATH_IMAGE040
The weight ratio of (A):
Figure 888057DEST_PATH_IMAGE041
=
Figure 489940DEST_PATH_IMAGE042
in the formula, the content of the active carbon is shown in the specification,
Figure 74505DEST_PATH_IMAGE043
for each vertebra edge high frequency pixel gray level
Figure 445443DEST_PATH_IMAGE039
And low frequency pixel gray scale
Figure 457261DEST_PATH_IMAGE040
The weight ratio of (a) to (b),
Figure 233193DEST_PATH_IMAGE044
entropy average of gray levels of low-frequency pixel points
Figure 621449DEST_PATH_IMAGE037
The ratio of the gray level of the high-frequency pixel point at the edge of each vertebra,
Figure 351288DEST_PATH_IMAGE045
for the frequency of occurrence of the grey levels,
Figure 330745DEST_PATH_IMAGE035
the frequency of the gray level of the high-frequency pixel points at the edge of the spine,
Figure 110483DEST_PATH_IMAGE036
the average frequency of the gray levels of the low-frequency pixel points is shown, and when the occurrence frequency of the gray levels of the high-frequency pixel points at the edge of the spine is equal to the average frequency of the gray levels of the low frequency, the gray levels can be kept as much as most of low-frequency information.
It should be noted that, first, for the high-frequency edge portion with the highest importance degree, the histogram equalization is the most ideal state when the frequency of occurrence of each gray level is equal. However, the total number of the pixels is fixed, and many low-distribution gray levels will inevitably disappear due to combination during histogram equalization, so that the gray levels combined in the low-frequency information do not need to be considered, and only the high-frequency gray levels are subjected to
Figure 36850DEST_PATH_IMAGE031
And
Figure 116802DEST_PATH_IMAGE029
the remaining processing is performed, only the frequency is adjusted, which inevitably causes the uncertainty of gray level combination, if the average frequency of the gray levels of the low-frequency pixels is too large, in order to make the average frequency larger
Figure 736002DEST_PATH_IMAGE035
=
Figure 534194DEST_PATH_IMAGE036
More low-profile gray levels are merged, which ignores the richness of the gray levels. The entropy is used for describing the disorder degree of the system, the larger the entropy is, the larger the uncertainty of the system is, so that the larger the image information amount is, and after each high-frequency low-distribution gray level which is easy to swallow is given a weight, the entropy value of each high-frequency low-distribution gray level is equal to the average value of the gray level entropy of the low-frequency pixel point, so that the high-frequency gray levels can be reserved.
The method for obtaining the gray levels of the low-frequency pixel points, the spine edge high-frequency pixel points and the non-spine edge high-frequency pixel points comprises the following steps:
(1) this embodiment is directed to non-spinal edge high frequency pixel
Figure 264252DEST_PATH_IMAGE029
Corresponding non-spine edge high frequency pixel gray scale
Figure 700175DEST_PATH_IMAGE046
Endowing a uniform weight value
Figure 490276DEST_PATH_IMAGE047
Calculating the weight mean value of the gray levels of all the vertebra edge high-frequency pixel points, and weighting the gray levels of the non-vertebra edge high-frequency pixel points
Figure 775764DEST_PATH_IMAGE046
And the weight mean value of the gray levels of all the high-frequency pixel points at the edge of the spine
Figure 43935DEST_PATH_IMAGE048
Is set to
Figure 98478DEST_PATH_IMAGE049
Because for
Figure 59481DEST_PATH_IMAGE029
It contains not only high-frequency information on the spine but also high-frequency information on muscle texture, organs, and the like. And therefore less important than the spine edge, a portion of the gray levels may be "sacrificed" to ensure preservation of the spine edge gray levels during histogram equalization.
(2) According to the weight ratio of the gray levels of all low-frequency pixel points and the gray level of each high-frequency pixel point at the edge of the vertebra, namely
Figure 832265DEST_PATH_IMAGE034
And
Figure 904126DEST_PATH_IMAGE032
the ratio, the proportional relation between the gray level weight of the non-spinal edge high-frequency pixel and the weight mean value of the gray levels of all spinal edge high-frequency pixels
Figure 311711DEST_PATH_IMAGE046
And
Figure 443616DEST_PATH_IMAGE048
the ratio of the gray scale weight of the high-frequency pixel at the spine edge to the gray scale weight of the low-frequency pixel and the gray scale of the high-frequency pixel at the non-spine edge can be obtained
Figure 703696DEST_PATH_IMAGE032
Figure 313668DEST_PATH_IMAGE046
Figure 342804DEST_PATH_IMAGE034
And carrying out weight distribution according to the proportional relation.
For example, the calculation method of the gray level weight of the spine edge high-frequency pixel point, the gray level weight of the low-frequency pixel point and the gray level weight of the non-spine edge high-frequency pixel point is as follows:
(1) suppose the gray level of the high-frequency pixel point at the edge of the vertebra is
Figure 380031DEST_PATH_IMAGE050
Figure 127407DEST_PATH_IMAGE051
The gray level of the non-vertebral edge high-frequency pixel point is
Figure 713296DEST_PATH_IMAGE052
Figure 393676DEST_PATH_IMAGE053
Low frequency pixel point gray scale of
Figure 867383DEST_PATH_IMAGE034
;
(2) Obtaining the weight ratio of the gray levels of all the low-frequency pixel points and the gray level of each high-frequency pixel point at the edge of the vertebra according to the stepThe weight ratio of the gray level of the point to the gray level of the high-frequency pixel point at the edge of each vertebra is
Figure 836476DEST_PATH_IMAGE054
,
Figure 788252DEST_PATH_IMAGE055
Then, then
Figure 791980DEST_PATH_IMAGE050
=,
Figure 374271DEST_PATH_IMAGE051
=
Figure 957430DEST_PATH_IMAGE056
;
(3) Obtaining the proportion of the gray level weight of the non-spine edge high-frequency pixel points to the weight average of the gray levels of all spine edge high-frequency pixel points according to the step:
Figure 712896DEST_PATH_IMAGE057
Figure 300013DEST_PATH_IMAGE058
wherein
Figure 912260DEST_PATH_IMAGE059
=
Figure 855945DEST_PATH_IMAGE060
Then, then
Figure 149523DEST_PATH_IMAGE061
Figure 862264DEST_PATH_IMAGE062
(4) Then obtain
Figure 848675DEST_PATH_IMAGE063
Obtaining the gray level of the high-frequency pixel point at the edge of the vertebra according to the proportional relation
Figure 279656DEST_PATH_IMAGE050
Figure 878390DEST_PATH_IMAGE051
High frequency gray scale of non-vertebral edge pixel
Figure 445637DEST_PATH_IMAGE052
Figure 602949DEST_PATH_IMAGE053
Gray level of low frequency pixel
Figure 521227DEST_PATH_IMAGE034
The weight of (c):
Figure 156607DEST_PATH_IMAGE050
the weight is
Figure 109520DEST_PATH_IMAGE064
Figure 437733DEST_PATH_IMAGE051
The weight is
Figure 878860DEST_PATH_IMAGE065
Figure 583510DEST_PATH_IMAGE052
The weight is
Figure 328613DEST_PATH_IMAGE066
Figure 827727DEST_PATH_IMAGE053
The weight is
Figure 455017DEST_PATH_IMAGE066
Figure 432201DEST_PATH_IMAGE034
The weight is
Figure 94126DEST_PATH_IMAGE067
Step five: and constructing an accumulation mapping function according to the gray level of the low-frequency pixel point, the gray level of the spine edge high-frequency pixel point and the weight of the gray level of the non-spine edge high-frequency pixel point, and performing histogram equalization on the whole spine image map by using the accumulation mapping function to obtain the processed whole spine image map.
The purpose of the step is to modify the mapping accumulation function of histogram equalization according to the gray level weight of the non-vertebra edge high-frequency pixel point, the gray level weight of the vertebra edge high-frequency pixel point and the gray level weight of the low-frequency pixel point, perform histogram equalization and enhance the image.
The cumulative mapping function after adding weight distribution during histogram equalization is as follows:
Figure 764142DEST_PATH_IMAGE068
in the formula, the first step is that,
Figure 380193DEST_PATH_IMAGE002
the weight of the gray level of the high-frequency pixel point at the edge of the vertebra,
Figure 161067DEST_PATH_IMAGE003
the weight of the gray level of the non-vertebral edge high-frequency pixel point,
Figure 677499DEST_PATH_IMAGE004
is the weight of the gray level of the low-frequency pixel,
Figure 518417DEST_PATH_IMAGE005
for the gray scale cumulative distribution function of the histogram equalized image,
Figure 120299DEST_PATH_IMAGE006
denotes the first
Figure 970443DEST_PATH_IMAGE007
The mapping function of the original image and the equalized image at each gray level,
Figure 341382DEST_PATH_IMAGE008
is shown in the original image
Figure 353200DEST_PATH_IMAGE009
The frequency with which the individual gray levels occur,
Figure 675335DEST_PATH_IMAGE010
the total number of the pixel points is,
Figure 63591DEST_PATH_IMAGE007
for grey levels of the original image
Figure 289036DEST_PATH_IMAGE011
Figure 675018DEST_PATH_IMAGE012
As is the number of grey levels in the original image,
Figure 251493DEST_PATH_IMAGE013
for the purpose of the normalized gray scale level,
Figure 361881DEST_PATH_IMAGE014
Figure 441833DEST_PATH_IMAGE015
,
Figure 61033DEST_PATH_IMAGE016
Figure 859225DEST_PATH_IMAGE017
respectively representing the gray levels of the original image as
Figure 323704DEST_PATH_IMAGE018
Figure 523741DEST_PATH_IMAGE019
Figure 812378DEST_PATH_IMAGE020
Figure 832287DEST_PATH_IMAGE007
The number of the pixel points.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (2)

1. A spine feature identification method based on image processing is characterized by comprising the following steps:
acquiring a full spine lateral image map, and performing semantic segmentation on the full spine lateral image map to obtain a full spine image map;
detecting the whole spine image map by using a sobel operator to obtain gradient edge pixel points of the whole spine image map;
carrying out frequency division filtering on the whole spine image map to obtain high-frequency pixel points and low-frequency pixel points of the whole spine image map;
determining spine edge high-frequency pixel points and non-spine edge high-frequency pixel points in the high-frequency pixel points by utilizing the gradient edge pixel points;
taking the gray levels corresponding to the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points as the gray levels of the spine edge high-frequency pixel points, the non-spine edge high-frequency pixel points and the low-frequency pixel points respectively;
obtaining a gray level histogram according to the gray level of the vertebra edge high-frequency pixel points, the gray level of the non-vertebra edge high-frequency pixel points and the frequency of the gray level of the low-frequency pixel points appearing in the whole vertebra image map;
acquiring the average frequency of the gray levels of all low-frequency pixel points in the gray histogram and the frequency of the gray level of each high-frequency pixel point at the edge of the vertebra;
obtaining the weight ratio of the gray levels of all the low-frequency pixel points and the gray level of each spine edge high-frequency pixel point according to the average frequency of the gray levels of all the low-frequency pixel points and the frequency of the gray level of each spine edge high-frequency pixel point;
the method for obtaining the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine comprises the following steps:
calculating the average frequency of the gray levels of all low-frequency pixel points in the gray histogram;
obtaining the average value of the entropies of the gray levels of all the low-frequency pixel points according to the average frequency of the gray levels of all the low-frequency pixel points;
acquiring the frequency of the gray level of each high-frequency pixel point at the edge of each vertebra;
taking the ratio of the frequency of the gray level of each spine edge high-frequency pixel point to the average value of the entropies of the gray levels of all low-frequency pixel points as the weight ratio of the gray level of each spine edge high-frequency pixel point to the gray level of each low-frequency pixel point;
determining the gray levels of all the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the weights of the gray levels of the non-spine edge high-frequency pixel points according to the weight ratio of the gray levels of all the low-frequency pixel points to the gray level of each spine edge high-frequency pixel point;
the method for determining the weight of the gray level of the low-frequency pixel point, the gray level of the high-frequency pixel point at the edge of the spine and the gray level of the high-frequency pixel point at the edge of the non-spine comprises the following steps:
calculating a weight average value of the gray levels of the high-frequency pixel points at the edge of the spine according to the weight ratio of the gray levels of the low-frequency pixel points to the gray level of each high-frequency pixel point at the edge of the spine;
setting the ratio of the weight mean value of the gray level of the high-frequency pixel points at the edge of the vertebra to the gray level weight of all the high-frequency pixel points at the edge of the non-vertebra as
Figure DEST_PATH_IMAGE001
Obtaining the proportional relation of the gray level of the low-frequency pixel point, the gray level of the spine edge high-frequency pixel point and the gray level of the non-spine edge high-frequency pixel point according to the weight ratio of the gray level of the low-frequency pixel point to the gray level of each spine edge high-frequency pixel point, the weight average value of the gray level of the spine edge high-frequency pixel point and the weight proportion of the gray levels of all non-spine edge high-frequency pixel points;
carrying out weight distribution according to the proportional relation of the gray level of the low-frequency pixel point, the gray level of the spine edge high-frequency pixel point and the gray level of the non-spine edge high-frequency pixel point;
constructing an accumulation mapping function according to the gray levels of the low-frequency pixel points, the gray levels of the spine edge high-frequency pixel points and the gray levels of the non-spine edge high-frequency pixel points, and performing histogram equalization on the whole spine image map by using the accumulation mapping function to obtain a processed whole spine image map;
the cumulative mapping function is as follows:
Figure 579403DEST_PATH_IMAGE002
in the formula,
Figure DEST_PATH_IMAGE003
the weight of the gray level of the high-frequency pixel point at the edge of the vertebra,
Figure 558860DEST_PATH_IMAGE004
the weight of the gray level of the non-vertebral edge high-frequency pixel point,
Figure DEST_PATH_IMAGE005
is the weight of the gray level of the low-frequency pixel,
Figure 902379DEST_PATH_IMAGE006
for the gray scale cumulative distribution function of the histogram equalized image,
Figure DEST_PATH_IMAGE007
denotes the first
Figure 359905DEST_PATH_IMAGE008
The mapping function of the original image and the equalized image at each gray level,
Figure DEST_PATH_IMAGE009
is shown in the original image
Figure 236595DEST_PATH_IMAGE010
The frequency with which the individual gray levels occur,
Figure DEST_PATH_IMAGE011
the total number of the pixel points is,
Figure 855795DEST_PATH_IMAGE008
for grey levels of the original image
Figure 653987DEST_PATH_IMAGE012
Figure DEST_PATH_IMAGE013
As is the number of grey levels in the original image,
Figure 413739DEST_PATH_IMAGE014
for the purpose of the normalized gray scale level,
Figure DEST_PATH_IMAGE015
Figure 144934DEST_PATH_IMAGE016
,
Figure DEST_PATH_IMAGE017
Figure 200615DEST_PATH_IMAGE018
respectively representing the gray levels of the original image as
Figure DEST_PATH_IMAGE019
Figure 228046DEST_PATH_IMAGE020
Figure DEST_PATH_IMAGE021
Figure 761796DEST_PATH_IMAGE008
The number of the pixel points.
2. The spine feature identification method based on image processing according to claim 1, wherein the method for determining spine edge high frequency pixel points and non-spine edge high frequency pixel points among the high frequency pixel points comprises:
taking gradient edge pixel points of the full spine image map as spine edge high-frequency pixel points in the high-frequency pixel points;
and other pixels except the high-frequency pixels at the edge of the spine in the high-frequency pixels are non-high-frequency pixels at the edge of the spine.
CN202210610910.6A 2022-06-01 2022-06-01 Spine feature identification method based on image processing Active CN114693682B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210610910.6A CN114693682B (en) 2022-06-01 2022-06-01 Spine feature identification method based on image processing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210610910.6A CN114693682B (en) 2022-06-01 2022-06-01 Spine feature identification method based on image processing

Publications (2)

Publication Number Publication Date
CN114693682A CN114693682A (en) 2022-07-01
CN114693682B true CN114693682B (en) 2022-08-26

Family

ID=82131266

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210610910.6A Active CN114693682B (en) 2022-06-01 2022-06-01 Spine feature identification method based on image processing

Country Status (1)

Country Link
CN (1) CN114693682B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115830459B (en) * 2023-02-14 2023-05-12 山东省国土空间生态修复中心(山东省地质灾害防治技术指导中心、山东省土地储备中心) Mountain forest grass life community damage degree detection method based on neural network
CN117745722B (en) * 2024-02-20 2024-04-30 北京大学 Medical health physical examination big data optimization enhancement method

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10336657A (en) * 1997-05-29 1998-12-18 Ricoh Co Ltd Image processor
JPH11136521A (en) * 1997-10-31 1999-05-21 Ricoh Co Ltd Picture data processor
CN106651818A (en) * 2016-11-07 2017-05-10 湖南源信光电科技有限公司 Improved Histogram equalization low-illumination image enhancement algorithm
CN108230260A (en) * 2017-12-06 2018-06-29 天津津航计算技术研究所 A kind of fusion method of new infrared image and twilight image
CN109919929A (en) * 2019-03-06 2019-06-21 电子科技大学 A kind of fissuring of tongue feature extracting method based on wavelet transformation
WO2020103601A1 (en) * 2018-11-21 2020-05-28 Zhejiang Dahua Technology Co., Ltd. Method and system for generating a fusion image
CN111899205A (en) * 2020-08-10 2020-11-06 国科天成(北京)科技有限公司 Image enhancement method of scene self-adaptive wide dynamic infrared thermal imaging

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110852977B (en) * 2019-10-29 2023-04-11 天津大学 Image enhancement method for fusing edge gray level histogram and human eye visual perception characteristics
CN114494256B (en) * 2022-04-14 2022-06-14 武汉金龙电线电缆有限公司 Electric wire production defect detection method based on image processing

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10336657A (en) * 1997-05-29 1998-12-18 Ricoh Co Ltd Image processor
JPH11136521A (en) * 1997-10-31 1999-05-21 Ricoh Co Ltd Picture data processor
CN106651818A (en) * 2016-11-07 2017-05-10 湖南源信光电科技有限公司 Improved Histogram equalization low-illumination image enhancement algorithm
CN108230260A (en) * 2017-12-06 2018-06-29 天津津航计算技术研究所 A kind of fusion method of new infrared image and twilight image
WO2020103601A1 (en) * 2018-11-21 2020-05-28 Zhejiang Dahua Technology Co., Ltd. Method and system for generating a fusion image
CN109919929A (en) * 2019-03-06 2019-06-21 电子科技大学 A kind of fissuring of tongue feature extracting method based on wavelet transformation
CN111899205A (en) * 2020-08-10 2020-11-06 国科天成(北京)科技有限公司 Image enhancement method of scene self-adaptive wide dynamic infrared thermal imaging

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
PET/CT肺部成像过程中的衰减校正和分割方法的研究;张倩,;《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》;20140215;第2014年卷(第2期);第I138-808页 *
Research on feature extraction algorithm of pavement disease;Lixia Cao 等,;《2021 International Conference on Electronic Information Engineering and Computer Science (EIECS)》;20211113;第2021卷;第1-4页 *
耦合边缘检测与优化的多尺度遥感图像融合法;谷志鹏 等,;《计算机工程与应用》;20170601;第53卷(第11期);第192-198页 *

Also Published As

Publication number Publication date
CN114693682A (en) 2022-07-01

Similar Documents

Publication Publication Date Title
CN114648530B (en) CT image processing method
CN114693682B (en) Spine feature identification method based on image processing
WO2021017297A1 (en) Artificial intelligence-based spine image processing method and related device
US7218763B2 (en) Method for automated window-level settings for magnetic resonance images
CN103026379B (en) The method calculating image noise level
CN112446880B (en) Image processing method, electronic device and readable storage medium
CN109410177A (en) A kind of image quality analysis method and system of super-resolution image
CN117237591A (en) Intelligent removal method for heart ultrasonic image artifacts
CN117876402B (en) Intelligent segmentation method for temporomandibular joint disorder image
CN111951215A (en) Image detection method and device and computer readable storage medium
CN104574392B (en) A kind of retrogression computer automatic grading method of interverbebral disc image
CN108038840B (en) Image processing method and device, image processing equipment and storage medium
CN117408988A (en) Artificial intelligence-based focus image analysis method and apparatus
CN118333893A (en) Preoperative blood vessel assessment method for free flap surgery
CN118229538A (en) Intelligent enhancement method for bone quality CT image
CN117237342B (en) Intelligent analysis method for respiratory rehabilitation CT image
CN114332255B (en) Medical image processing method and device
CN116029934A (en) Low-dose DR image and CT image denoising method
Yang et al. Fusion of CT and MR images using an improved wavelet based method
Kumar et al. Semiautomatic method for segmenting pedicles in vertebral radiographs
CN115205241A (en) Metering method and system for apparent cell density
CN110084770B (en) Brain image fusion method based on two-dimensional Littlewood-Paley empirical wavelet transform
Pancholi et al. A Review of Noise Reduction Filtering Techniques for MRI Images
CN114418920B (en) Endoscope multi-focus image fusion method
CN118570205B (en) Image processing method, device and system based on portable X-ray imaging

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant