CN103475881B - The image JND threshold value computational methods of view-based access control model attention mechanism in DCT domain - Google Patents

The image JND threshold value computational methods of view-based access control model attention mechanism in DCT domain Download PDF

Info

Publication number
CN103475881B
CN103475881B CN201310413594.4A CN201310413594A CN103475881B CN 103475881 B CN103475881 B CN 103475881B CN 201310413594 A CN201310413594 A CN 201310413594A CN 103475881 B CN103475881 B CN 103475881B
Authority
CN
China
Prior art keywords
image
block
saliency
salient
jnd
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201310413594.4A
Other languages
Chinese (zh)
Other versions
CN103475881A (en
Inventor
张冬冬
高利晶
臧笛
孙杳如
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tongji University
Original Assignee
Tongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tongji University filed Critical Tongji University
Priority to CN201310413594.4A priority Critical patent/CN103475881B/en
Publication of CN103475881A publication Critical patent/CN103475881A/en
Application granted granted Critical
Publication of CN103475881B publication Critical patent/CN103475881B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

一种DCT域内基于视觉注意力机制的图像JND阈值计算方法。本发明提出了两种将显著度和块分类相结合的方案,一种是用单个点的视觉注意力掩蔽因子和块分类的掩蔽因子按照点对点的方式相结合,另一种是用每个块的平均显著度代表整个块的显著度,然后将基于每个块的视觉注意力掩蔽因子和块分类的掩蔽因子按照块到块的方式相结合。使用综合的对比度掩蔽函数计算得到的值对传统的JND阈值进行调制,最终得到更加准确的JND阈值。两种方法都能有效地提高JND阈值的准确度,从而使得JND阈值和人眼视觉系统更加匹配。本发明提出的图像JND阈值计算方法实现的模型可以容纳更多的噪声,在PSNR方面,模型平均可以提高0.54DB。

A visual attention mechanism based image JND threshold calculation method in the DCT domain. The present invention proposes two schemes combining saliency and block classification, one is to combine the visual attention masking factor of a single point with the masking factor of block classification in a point-to-point manner, and the other is to use each block The average saliency of represents the saliency of the whole block, and then combines the visual attention masking factor based on each block and the masking factor of the block classification in a block-to-block manner. The traditional JND threshold is modulated using the value calculated by the integrated contrast masking function, resulting in a more accurate JND threshold. Both methods can effectively improve the accuracy of the JND threshold, thus making the JND threshold more closely match the human visual system. The model realized by the image JND threshold calculation method proposed by the present invention can accommodate more noise, and in terms of PSNR, the average model can be improved by 0.54DB.

Description

DCT域内基于视觉注意力机制的图像JND阈值计算方法Image JND Threshold Calculation Method Based on Visual Attention Mechanism in DCT Domain

技术领域technical field

本发明涉及图像/视频编码技术领域。The invention relates to the technical field of image/video coding.

技术背景technical background

传统的图像/视频编码技术主要针对空间域冗余、时间域冗余以及统计冗余进行压缩编码,但很少考虑到人眼视觉系统特性和心理效应,因此大量视觉冗余数据被编码并传输,为了进一步提高编码的效率,研究人员开始了致力于去除视觉冗余的研究。目前一个表征视觉冗余的有效方法就是基于心理学和生理学的最小可察觉失真模型,简称JND模型,也可称为恰可察觉失真模型,即人眼不能感知的变化,由于人眼的各种屏蔽效应,人眼只能觉察超过某一阈值的噪声,该阈值就是人眼的恰可觉察失真,代表着图像中的视觉冗余度。JND模型常用来指导图像或视频的感知编码和处理,如预处理、自适应量化、码流控制、运动估计等。Traditional image/video coding technology mainly compresses and codes spatial redundancy, time domain redundancy, and statistical redundancy, but rarely considers the characteristics and psychological effects of the human visual system, so a large amount of visual redundant data is encoded and transmitted , in order to further improve the efficiency of coding, researchers have started research on removing visual redundancy. At present, an effective method to represent visual redundancy is the least detectable distortion model based on psychology and physiology, referred to as the JND model, which can also be called the just detectable distortion model, that is, the changes that the human eye cannot perceive, due to various human eyes. The shielding effect means that the human eye can only perceive noise above a certain threshold, which is the just detectable distortion of the human eye, representing the visual redundancy in the image. The JND model is often used to guide the perceptual coding and processing of images or videos, such as preprocessing, adaptive quantization, stream control, motion estimation, etc.

现有的可察觉失真(JND)模型可以大致分为两类:第一类为像素域JND模型,其基本原理大多是通过表征亮度自适应效应和纹理掩蔽效应来建模,例如文献1(参见X.Yang,W.Lin,Z.Lu,E.P.Ong,and S.Yao,“Just-noticeable-distortion profile withnonlinear additivity model for perceptual masking color images”,IEEETrans.Circuits Syst.Video Technol.,vol.15,no.6,pp742-752,Jun.2005)中提出了基于空域的彩色图像JND模型,但是由于其无法很好的将对比敏感度函数(ContrastSensitive Function,CSF)整合进来,因此这类模型没有办法得到精确的JND值,常作为计算JND阈值的快速方法来使用。The existing perceivable distortion (JND) models can be roughly divided into two categories: the first category is the pixel-domain JND model, and its basic principles are mostly modeled by characterizing the brightness adaptive effect and the texture masking effect, such as literature 1 (see X.Yang, W.Lin, Z.Lu, E.P.Ong, and S.Yao, "Just-noticeable-distortion profile with nonlinear additivity model for perceptual masking color images", IEEETrans.Circuits Syst.Video Technol., vol.15, no.6, pp742-752, Jun.2005) proposed a color image JND model based on airspace, but because it cannot integrate the contrast sensitivity function (ContrastSensitive Function, CSF) well, there is no way for this type of model Get an accurate JND value, often used as a quick way to calculate the JND threshold.

第二类JND模型为子带JND模型,这类模型是在变换域中进行计算,例如DCT域、小波域、CONTOURLET域等。由于大多数图像/视频编码标准都是基于DCT域(如JPEG、H.261/3/4、MPEG-1/2/4),因此基于DCT域的JND模型得到了很多研究者的关注,例如文献2(参见Z.Wei and K.N.Ngan,“Spatial just noticeable distortion profile for image inDCT domain,”In Proc.IEEE Int.Conf.Multimeda and Expo,pp.925-928,2008.)中结合图像的亮度自适应特性,空间对比度效应以及基于块分类的对比度掩盖效应,但是该模型并没有考虑人眼的视觉注意力机制对JND模型的影响,因此计算精度有待进一步提高。The second type of JND model is the subband JND model, which is calculated in the transform domain, such as DCT domain, wavelet domain, CONTOURLET domain and so on. Since most image/video coding standards are based on the DCT domain (such as JPEG, H.261/3/4, MPEG-1/2/4), the JND model based on the DCT domain has attracted the attention of many researchers, such as Document 2 (see Z.Wei and K.N.Ngan, "Spatial just noticeable distortion profile for image inDCT domain," In Proc.IEEE Int.Conf.Multimeda and Expo, pp.925-928, 2008.) combined with the brightness of the image from Adaptive characteristics, spatial contrast effect and contrast masking effect based on block classification, but this model does not consider the influence of the human eye's visual attention mechanism on the JND model, so the calculation accuracy needs to be further improved.

发明内容Contents of the invention

在WEI的模型的基础上本发明结合视觉注意力机制提出了一个新的DCT域内图像JND模型建模方法,通过综合考虑视觉注意力效应和对比度掩蔽效应设计出了一种综合的调制函数和亮度自适应效应一起对空间对比度敏感函数进行调制的方法。On the basis of the WEI model, the present invention combines the visual attention mechanism to propose a new JND model modeling method for images in the DCT domain, and designs a comprehensive modulation function and brightness by comprehensively considering the visual attention effect and contrast masking effect A method for modulating spatial contrast-sensitive functions together with adaptive effects.

为此,本发明给出技术方案实施步骤为:For this reason, the present invention provides technical scheme implementation steps as:

一种基于DCT域的图像可觉察失真度计算方法,采用如下技术方案,包括以下步骤:A DCT domain-based image perceivable distortion calculation method adopts the following technical solution, including the following steps:

步骤S1:将选定的图像进行8x8的DCT变换,将其由空域变换到DCT域。Step S1: Perform 8x8 DCT transformation on the selected image, and transform it from the spatial domain to the DCT domain.

步骤S2:在DCT域,根据空间对比度效应阈值和亮度自适应调制因子的乘积计算获得的基本的可觉察失真度JND值。Step S2: In the DCT domain, calculate the basic perceivable distortion degree JND value obtained according to the product of the spatial contrast effect threshold and the brightness adaptive modulation factor.

步骤S3:利用canny边缘检测器对图像进行分块,分为平滑块、边缘块和纹理块,得到基于块结构的对比度掩蔽因子。Step S3: using the canny edge detector to divide the image into smooth blocks, edge blocks and texture blocks to obtain a contrast masking factor based on the block structure.

步骤S4:利用视觉注意模型对图像进行显著性检测,得到图像的显著图。Step S4: Use the visual attention model to perform saliency detection on the image to obtain a saliency map of the image.

步骤S5:按照步骤S4所得的显著图对图像进行分割,将图像分为显著区域和非显著区域,然后基于每个点的显著值,得到基于视觉注意力机制的对比度掩蔽因子。Step S5: Segment the image according to the saliency map obtained in step S4, divide the image into salient regions and non-salient regions, and then obtain the contrast masking factor based on the visual attention mechanism based on the saliency value of each point.

或者步骤S5为:首先将图像分割成显著区域和非显著区域,然后对步骤S4所得的显著图分块,用每个块的显著值的平均值代替整个块的显著值,并基于每个块的显著值得到基于视觉注意力机制的对比度掩蔽因子。Or step S5 is: first divide the image into salient areas and non-salient areas, then divide the salient map obtained in step S4 into blocks, replace the salient value of the entire block with the average value of the salient values of each block, and based on each block The saliency value of is obtained as a contrast masking factor based on the visual attention mechanism.

步骤S6:将步骤S5得到的图像显著区域和非显著区域的分割结果和步骤S3所得的块分类结果相结合,对图像进行更加细致的分块,并将基于视觉注意力机制的对比度掩蔽因子和基于块结构的对比度掩蔽因子按照线性关系结合起来,得到综合的对比度掩蔽调制函数,Step S6: Combining the segmentation results of salient areas and non-salient areas of the image obtained in step S5 with the block classification results obtained in step S3, the image is divided into more detailed blocks, and the contrast masking factor based on the visual attention mechanism and The contrast masking factors based on the block structure are combined according to a linear relationship to obtain a comprehensive contrast masking modulation function,

步骤S7:将步骤S6计算得到的调制函数值对步骤S2计算得到的JND基本阈值进行调制,得到最终的JND阈值。Step S7: Modulate the modulation function value calculated in step S6 to the basic JND threshold calculated in step S2 to obtain a final JND threshold.

上述技术方案体现出的关键技术要点:The key technical points reflected in the above technical solutions:

1、针对传统图像恰可察觉失真模型没有考虑视觉注意力机制这个问题,本发明提出了两种DCT域内基于视觉注意力机制的图像恰可察觉失真模型的建模算法,一种是,通过计算基于图像显著度的视觉注意力调制因子,并将该基于像素点的视觉注意力掩蔽因子和基于块结构的对比度掩蔽因子相结合,计算得到综合的对比度掩蔽函数,对传统的基于空间对比度效应和亮度自适应效应的JND阈值进行调制;第二种是,通过计算图像的显著图,并用每个块的显著度平均值代替整个块的显著度因子,然后建立基于块的视觉注意力掩蔽因子和基于块结构的对比度掩蔽因子相结合,得到综合的对比度掩蔽函数对基本的JND阈值进行调制。两种方法都能有效地提高JND阈值的准确度,从而使得JND阈值和人眼视觉系统更加匹配。1. In view of the problem that the traditional image just perceptible distortion model does not consider the visual attention mechanism, the present invention proposes two modeling algorithms for the image just perceptible distortion model based on the visual attention mechanism in the DCT domain. One is, by calculating The visual attention modulation factor based on image saliency, and the pixel-based visual attention masking factor and the block structure-based contrast masking factor are combined to calculate a comprehensive contrast masking function, which is based on the traditional spatial contrast effect and The JND threshold of the brightness adaptive effect is modulated; the second is, by calculating the saliency map of the image, and replacing the saliency factor of the whole block with the saliency average value of each block, and then establishing a block-based visual attention masking factor and The block-structure-based contrast masking factors are combined to obtain a synthetic contrast masking function that modulates the underlying JND threshold. Both methods can effectively improve the accuracy of the JND threshold, thus making the JND threshold more closely match the human visual system.

2、本发明提出了视觉注意力掩蔽效应的概念,通过视觉显著度模拟人眼对图像像素点的关注度,从而建立基于视觉显著度的视觉注意力掩蔽效应因子。2. The present invention proposes the concept of visual attention masking effect, and simulates the human eye's attention to image pixels through visual salience, thereby establishing a visual attention masking effect factor based on visual salience.

3、本发明提出了两种将显著度和块分类相结合的方案,一种是用单个点的视觉注意力掩蔽因子和块分类的掩蔽因子按照点对点的方式相结合,另一种是用每个块的平均显著度代表整个块的显著度,然后将基于每个块的视觉注意力掩蔽因子和块分类的掩蔽因子按照块到块的方式相结合。3. The present invention proposes two schemes combining saliency and block classification, one is to combine the masking factor of visual attention of a single point and the masking factor of block classification in a point-to-point manner, and the other is to use each The average saliency of each block represents the saliency of the whole block, and then combines the visual attention masking factor based on each block and the masking factor of the block classification in a block-to-block manner.

4、通过将视觉显著度和块分类相结合,对图像进行更加精细和准确的分块。4. By combining visual saliency and block classification, images are segmented more finely and accurately.

5、基于每个块的显著特性和块结构特性,通过设置不同的显著度调制因子和块结构调制因子,从而将视觉注意力掩蔽因子和块分类对比度掩蔽因子结合起来,得到综合的对比度掩蔽函数。5. Based on the saliency and block structure characteristics of each block, by setting different saliency modulation factors and block structure modulation factors, the visual attention masking factor and the block classification contrast masking factor are combined to obtain a comprehensive contrast masking function .

本发明方法的有益效果为:使用综合的对比度掩蔽函数计算得到的值对传统的JND阈值进行调制,最终得到更加准确的JND阈值。在保证同样的视觉主观质量的前提下,本发明提出的图像JND阈值计算方法实现的模型可以容纳更多的噪声。The beneficial effect of the method of the present invention is that the traditional JND threshold is modulated by using the value calculated by the comprehensive contrast masking function, and finally a more accurate JND threshold is obtained. Under the premise of ensuring the same visual subjective quality, the model realized by the image JND threshold calculation method proposed by the present invention can accommodate more noise.

附图说明Description of drawings

图1是本发明DCT域的基于的基于视觉注意力机制的图像可觉察失真度模型框图。Fig. 1 is a block diagram of an image perceivable distortion model based on a visual attention mechanism in the DCT domain of the present invention.

图2是本发明实例Airplane图像。Fig. 2 is the Airplane image of the example of the present invention.

图3是本发明实例Airplane图像进行块分类之后的图像。Fig. 3 is an image after block classification of the Airplane image of the example of the present invention.

图4是本发明实例Airplane图像进行显著度分类之后的图像。Fig. 4 is the image after the saliency classification of the Airplane image of the example of the present invention.

图5是本发明实例Airplane图像综合考虑显著度和块结构的细致分块结果。Fig. 5 is the detailed block result of the Airplane image of the example of the present invention considering the saliency and the block structure comprehensively.

图6为本发明DCT域内基于视觉注意力机制的图像JND阈值计算方法流程图。FIG. 6 is a flowchart of a method for calculating an image JND threshold based on a visual attention mechanism in the DCT domain of the present invention.

具体实施方式detailed description

下面以具体实例结合附图对本发明作进一步说明:Below in conjunction with accompanying drawing, the present invention will be further described with specific example:

本发明提供的实例采用MATLAB7作为仿真实验平台,以512×512的bmp灰度图像Airplane作为选定的测试图像,下面结合每个步骤详细描述本实例:The example that the present invention provides adopts MATLAB7 as the emulation experiment platform, uses the bmp grayscale image Airplane of 512 * 512 as the selected test image, describes this example in detail below in conjunction with each step:

步骤(1),选定512×512的bmp灰度图像作为输入测试的图像,将其进行8×8的DCT变换,将其由空域变换到DCT域;Step (1), select the bmp grayscale image of 512 * 512 as the image of input test, carry out the DCT transformation of 8 * 8 to it, transform it into DCT domain by space domain;

步骤(2),在DCT域,根据空间对比度基本阈值和亮度自适应调制因子的乘积计算获得可觉察失真JND值,其计算公式如下:In step (2), in the DCT domain, the perceivable distortion JND value is obtained by calculating the product of the basic threshold value of the spatial contrast and the adaptive modulation factor of the brightness, and the calculation formula is as follows:

TJND(n,i,j)=TBasic(n,i,j)×Flum(n) (1)T JND (n,i,j)=T Basic (n,i,j)×F lum (n) (1)

其中,TBasic(n,i,j)代表空间对比度敏感阈值,Flum(n)代表亮度自适应调整因子,n为DCT块的索引,W为图像的宽度,H为图像的高度,i,j为DCT块中系数的索引,1≤i≤64,1≤j≤64。一般的DCT块都为8x8大小,那么该测试图像就分为个DCT块,n取1到4096之间数值,i、j取1到64之间数值。Among them, T Basic (n,i,j) represents the spatial contrast sensitivity threshold, F lum (n) represents the brightness adaptive adjustment factor, n is the index of the DCT block, W is the width of the image, H is the height of the image, i, j are the indices of the coefficients in the DCT block, 1≤i≤64, 1≤j≤64. The general DCT block is 8x8 in size, then the test image is divided into A DCT block, n takes a value between 1 and 4096, and i and j take a value between 1 and 64.

上式(l)中的基本阈值TBasic(n,i,j)采用如下方法计算获得:The basic threshold T Basic (n, i, j) in the above formula (l) is calculated by the following method:

DCT子带的频率可以表示如下:The frequencies of the DCT subbands can be expressed as follows:

ωω ii ,, jj == 11 22 NN (( ii // θθ xx )) 22 ++ (( jj // θθ ythe y )) 22 -- -- -- (( 22 ))

其中θx=θy=2·arctan(γ/2·l)是一个像素的水平和垂直视角,l是图像的观察距离,在该发明中,l为图像宽度的3倍,γ代表显示器上一个像素显示的宽度或者长度。由上所述,DCT块的空间对比度敏感阈值为:Wherein θ x = θ y = 2·arctan(γ/2·l) is the horizontal and vertical viewing angle of a pixel, l is the viewing distance of the image, in this invention, l is 3 times of the image width, and γ represents the One pixel display width or length. From the above, the spatial contrast sensitivity threshold of the DCT block is:

式(3)中:s代表空间集合效应,在该发明中取0.25,代表倾斜效应,其中r=0.6,Фi、Фj是DCT归一化系数,代表相应DCT系数的方向角:In the formula (3): s represents the spatial set effect, gets 0.25 in this invention, Represents the tilt effect, where r=0.6, Ф i and Ф j are DCT normalization coefficients, Represents the orientation angle of the corresponding DCT coefficient:

ΦΦ hh == 11 // NN mm == 00 22 // NN mm >> 00 -- -- -- (( 44 ))

参数a=l.33,b=0.11,c=0.005。Parameters a=1.33, b=0.11, c=0.005.

上式(1)中的亮度自适应调制因子Flum(n)采用如下方法计算获得:The brightness adaptive modulation factor F lum (n) in the above formula (1) is calculated by the following method:

Ff lulu nno (( nno )) == (( 6060 -- II )) // 150150 ++ 11 II &le;&le; 6060 11 6060 << II << 170170 (( II -- 170170 )) // 425425 ++ 11 II &GreaterEqual;&Greater Equal; 170170 -- -- -- (( 66 ))

式(6)中,I代表每个8×8块大的平均亮度值。In formula (6), I represents the average brightness value of each 8×8 block.

然后按照公式(1)计算得到传统的JND阈值。Then calculate according to the formula (1) to obtain the traditional JND threshold.

步骤(3),对图像进行块分类,计算基于块分类的对比度掩蔽因子Fcontrast(n,i,j)。In step (3), block classification is performed on the image, and a contrast masking factor F contrast (n,i,j) based on block classification is calculated.

将图像中的所有8×8大小DCT块进行分类,分成纹理、平坦和边缘三类(如图2所示);用canny算子来检测计算图像中的边缘信息,通过计算每个8×8块中的边缘像素密度ρedge来对块进行分类,具体过程如下:Classify all 8×8 DCT blocks in the image into three categories: texture, flat and edge (as shown in Figure 2); use the canny operator to detect and calculate the edge information in the image, by calculating each 8×8 The edge pixel density ρ edge in the block is used to classify the block, and the specific process is as follows:

ρedge=(∑8X8edge)/N2 (7)ρ edge =(∑ 8X8 edge)/N 2 (7)

其中,ρedge表示每个8X8块中通过canny算子得到的边缘像素的总数;N为DCT块大小,在本实例中取N=8;Wherein, ρ edge represents the total number of edge pixels obtained by the canny operator in each 8×8 block; N is the DCT block size, and N=8 is taken in this example;

块的分类方法为:The block classification methods are:

由8×8的DCT块类型可知,块间掩蔽效应因子为:According to the 8×8 DCT block type, the inter-block masking effect factor is:

又考虑到相邻子带间的掩蔽效应,传统的基于块分类的对比度掩蔽函数表达式如下:Considering the masking effect between adjacent subbands, the traditional block classification-based contrast masking function is expressed as follows:

步骤(4),对图像进行显著性检测,得到图像的显著图。Step (4), performing saliency detection on the image to obtain a saliency map of the image.

在本实例中利用谱残差方法对图像进行显著性检测:In this example, the spectral residual method is used to detect the saliency of the image:

其中,P(f)和R(f)分别代表图像傅里叶变换之后的相位谱和幅度谱,代表傅里叶逆变换,g(σ)代表高斯滤波器,σ是滤波器窗口大小,在本实例中,σ取图像宽度的0.32倍。S(x)代表每个图像的显著图。Among them, P(f) and R(f) represent the phase spectrum and amplitude spectrum after Fourier transform of the image respectively, Represents the inverse Fourier transform, g(σ) represents the Gaussian filter, σ is the filter window size, in this example, σ is 0.32 times the image width. S(x) represents the saliency map of each image.

步骤(5),若选择方案一,则参照(a)描述,若选择方案二,则参照(b)描述:Step (5), if option 1 is selected, refer to (a) description, if option 2 is selected, refer to (b) description:

(a)基于图像的显著图,计算图像的显著度对比度掩蔽因子:(a) Based on the saliency map of the image, calculate the saliency contrast masking factor of the image:

Fvs(n,i,j)=μ·(Smax-S(n,i,j)) (12)F vs (n,i,j)=μ·(S max -S(n,i,j)) (12)

图像的显著度值越大说明人眼对该点的关注度越高,其中Smax代表显著图归一化之后的最大值,S(n,i,j)代表显著图归一化之后每个块中每个点对应位置(i,j)的显著度。μ代表调制因子,在本实例中取μ=1.0。The larger the saliency value of the image, the higher the attention of the human eye to this point, where S max represents the maximum value after normalization of the saliency map, and S(n,i,j) represents each Each point in the block corresponds to the saliency of position (i,j). μ represents the modulation factor, and μ=1.0 is taken in this example.

(b)基于图像的显著图,用每个块的显著度平均值代表整个块的显著度。并基于每个块的显著度,计算图像的显著度对比度掩蔽因子:(b) Image-based saliency map, where the average saliency of each block is used to represent the saliency of the whole block. And based on the saliency of each block, calculate the saliency contrast masking factor of the image:

Fvs(n)=μ·(Smax(n)-S(n)) (13)F vs (n)=μ·(S max (n)-S(n)) (13)

其中,Smax(n)代表所有块的显著度中的最大值,S(n)代表每个块的显著度。μ代表调制因子,在本实例中取μ=1.0。Fvs(n)代表每个块的显著度掩蔽因子。Among them, S max (n) represents the maximum value of the saliency of all blocks, and S(n) represents the saliency of each block. μ represents the modulation factor, and μ=1.0 is taken in this example. F vs (n) represents the saliency masking factor for each block.

步骤(6),从块结构角度和显著度角度两方面综合出发,对图像进行更加细致的分类。In step (6), the images are classified in more detail from both the perspective of block structure and the perspective of saliency.

首先按照图像的显著度将图像分为显著区和非显著区域(如图3所示):First, according to the saliency of the image, the image is divided into salient areas and non-salient areas (as shown in Figure 3):

其中T代表分割阈值,在本实例中T是显著图的均值。where T represents the segmentation threshold, in this example T is the mean value of the saliency map.

结合显著度分割结果和块分类结果,对图像进行更细致的分类,如图4所示。Combining the results of saliency segmentation and block classification, the image is classified more finely, as shown in Figure 4.

步骤(7),针对不同的块结构,设置不同的显著度调整因子和块结构调整因子,从而计算综合的调制函数:Step (7), according to different block structures, set different salience adjustment factors and block structure adjustment factors, so as to calculate the comprehensive modulation function:

Ff cc oo nno tt rr aa sthe s tt vv sthe s (( nno ,, ii ,, jj )) == &alpha;&alpha; &times;&times; Ff vv sthe s ++ &beta;&beta; &times;&times; Ff cc oo nno tt rr aa sthe s tt (( nno ,, ii ,, jj )) -- -- -- (( 1515 ))

式(15)中,若采用方案一,则Fvs代表使用式(12)计算所得的显著度掩蔽因子,若采用方案二,则Fvs代表使用式(13)计算所得的显著度掩蔽因子。Fcontrast(n,i,j)代表式(10)计算得到的基于块结构的对比度掩蔽因子。显著度调制因子α和块结构调整因子β的取值基于不同的块结构:In formula (15), if scheme 1 is adopted, F vs represents the saliency masking factor calculated using formula (12); if scheme 2 is adopted, F vs represents the saliency masking factor calculated using formula (13). F contrast (n, i, j) represents the contrast masking factor based on the block structure calculated by formula (10). The values of saliency modulation factor α and block structure adjustment factor β are based on different block structures:

步骤(8),基于步骤(2)计算得到的JND基本阈值,利用步骤(7)计算得到综合的调制函数,对JND基本阈值进行调制:Step (8), based on the JND basic threshold calculated in step (2), use step (7) to calculate and obtain a comprehensive modulation function, and modulate the JND basic threshold:

TT JJ NN DD. (( nno ,, ii ,, jj )) == TT BB aa sthe s ii cc (( nno ,, ii ,, jj )) &times;&times; Ff ll uu mm (( nno )) &times;&times; Ff cc oo nno tt rr aa sthe s tt vv sthe s (( nno ,, ii ,, jj )) -- -- -- (( 1717 ))

综合以上所有步骤计算得到图像的JND阈值,该阈值综合考虑了空间对比度效应,亮度自适应效应,块分类对比度掩蔽效应以及视觉注意力机制,所以该阈值和人眼的视觉系统更加吻合,更加精确。The JND threshold of the image is calculated by combining all the above steps. The threshold comprehensively considers the spatial contrast effect, the brightness adaptive effect, the block classification contrast masking effect and the visual attention mechanism, so the threshold is more consistent with the human visual system and more accurate. .

本发明的创新点:Innovation point of the present invention:

提出了将图像显著度应用到改进图像恰可察觉失真模型的思路。 The idea of applying image saliency to improve the just perceptible distortion model of image is proposed.

提出了基于图像显著度的视觉注意力掩蔽因子,该因子描述了人眼对图像的关注程度。 A visual attention masking factor based on image saliency is proposed, which describes the degree of human eye's attention to the image.

综合考虑图像的显著性特征和块结构特征,对图像进行更加准确和细致的分类。 Considering the salient feature and block structure feature of the image comprehensively, the image can be classified more accurately and meticulously.

基于不同块的显著特性和块结构特性综合考虑,构造了综合的调制函数对传统的JND模型进行调制,计算得到更加吻合视觉系统的JND阈值。 Based on the comprehensive consideration of the salient characteristics and block structure characteristics of different blocks, a comprehensive modulation function is constructed to modulate the traditional JND model, and the JND threshold value that is more consistent with the visual system is calculated.

Claims (1)

1.一种基于DCT域的图像可觉察失真度计算方法,包括以下步骤:1. An image perceptible distortion calculation method based on the DCT domain, comprising the following steps: 步骤S1:将选定的图像进行8x8的DCT变换,将其由空域变换到DCT域;Step S1: Perform 8x8 DCT transformation on the selected image, and transform it from the spatial domain to the DCT domain; 步骤S2:在DCT域,根据空间对比度效应阈值和亮度自适应调制因子的乘积计算获得的基本的可觉察失真度JND值;Step S2: In the DCT domain, calculate the basic perceptible distortion degree JND value obtained according to the product of the spatial contrast effect threshold and the brightness adaptive modulation factor; 步骤S3:利用canny边缘检测器对图像进行分块,分为平滑块、边缘块和纹理块,得到基于块结构的对比度掩蔽因子;Step S3: using the canny edge detector to divide the image into smooth blocks, edge blocks and texture blocks to obtain a contrast masking factor based on the block structure; 步骤S4:利用视觉注意模型对图像进行显著性检测,得到图像的显著图;Step S4: using the visual attention model to perform saliency detection on the image to obtain a saliency map of the image; 步骤S5:Step S5: 按照步骤S4所得的显著图对图像进行分割,将图像分为显著区域和非显著区域,然后基于每个点的显著值,得到基于视觉注意力机制的对比度掩蔽因子;Segment the image according to the saliency map obtained in step S4, divide the image into salient areas and non-salient areas, and then obtain a contrast masking factor based on the visual attention mechanism based on the salient value of each point; 或者步骤S5为:首先将图像分割成显著区域和非显著区域,然后对步骤S4所得的显著图分块,用每个块的显著值的平均值代替整个块的显著值,并基于每个块的显著值得到基于视觉注意力机制的对比度掩蔽因子;Or step S5 is: first divide the image into salient areas and non-salient areas, then divide the salient map obtained in step S4 into blocks, replace the salient value of the entire block with the average value of the salient values of each block, and based on each block The saliency value of is obtained as a contrast masking factor based on the visual attention mechanism; 步骤S6:将步骤S5得到的图像显著区域和非显著区域的分割结果和步骤S3所得的块分类结果相结合,对图像进行更加细致的分块,并将基于视觉注意力机制的对比度掩蔽因子和基于块结构的对比度掩蔽因子按照线性关系结合起来,得到综合的对比度掩蔽调制函数;Step S6: Combining the segmentation results of salient areas and non-salient areas of the image obtained in step S5 with the block classification results obtained in step S3, the image is divided into more detailed blocks, and the contrast masking factor based on the visual attention mechanism and The contrast masking factors based on the block structure are combined according to a linear relationship to obtain a comprehensive contrast masking modulation function; 步骤S7:将步骤S6计算得到的对比度掩蔽调制函数值对步骤S2计算得到的JND基本阈值进行调制,得到最终的JND阈值。Step S7: Modulate the JND basic threshold calculated in step S2 with the contrast masking modulation function value calculated in step S6 to obtain a final JND threshold.
CN201310413594.4A 2013-09-12 2013-09-12 The image JND threshold value computational methods of view-based access control model attention mechanism in DCT domain Expired - Fee Related CN103475881B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310413594.4A CN103475881B (en) 2013-09-12 2013-09-12 The image JND threshold value computational methods of view-based access control model attention mechanism in DCT domain

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310413594.4A CN103475881B (en) 2013-09-12 2013-09-12 The image JND threshold value computational methods of view-based access control model attention mechanism in DCT domain

Publications (2)

Publication Number Publication Date
CN103475881A CN103475881A (en) 2013-12-25
CN103475881B true CN103475881B (en) 2016-11-23

Family

ID=49800559

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310413594.4A Expired - Fee Related CN103475881B (en) 2013-09-12 2013-09-12 The image JND threshold value computational methods of view-based access control model attention mechanism in DCT domain

Country Status (1)

Country Link
CN (1) CN103475881B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104219525B (en) * 2014-09-01 2017-07-18 国家广播电影电视总局广播科学研究院 Perception method for video coding based on conspicuousness and minimum discernable distortion
CN105491391A (en) * 2014-09-15 2016-04-13 联想(北京)有限公司 Image compression method and electronic equipment
CN104754320B (en) * 2015-03-27 2017-05-31 同济大学 A kind of 3D JND threshold values computational methods
CN109525847B (en) * 2018-11-13 2021-04-30 华侨大学 Just noticeable distortion model threshold calculation method
CN109948699B (en) * 2019-03-19 2020-05-15 北京字节跳动网络技术有限公司 Method and device for generating feature map
CN109948700B (en) * 2019-03-19 2020-07-24 北京字节跳动网络技术有限公司 Method and device for generating feature map
CN109902763B (en) * 2019-03-19 2020-05-15 北京字节跳动网络技术有限公司 Method and device for generating feature map
CN110251076B (en) * 2019-06-21 2021-10-22 安徽大学 Contrast-based saliency detection method and device for fusion visual attention
CN112437302B (en) * 2020-11-12 2022-09-13 深圳大学 JND prediction method and device for screen content image, computer device and storage medium
CN112435188B (en) * 2020-11-23 2023-09-22 深圳大学 JND prediction method and device based on direction weight, computer equipment and storage medium
CN114666619B (en) * 2022-03-11 2024-05-03 平安国际智慧城市科技股份有限公司 Watermarking method, device and equipment for video file and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101621708A (en) * 2009-07-29 2010-01-06 武汉大学 Method for computing perceptible distortion of color image based on DCT field
CN101710995A (en) * 2009-12-10 2010-05-19 武汉大学 Video coding system based on vision characteristic
CN102750706A (en) * 2012-07-13 2012-10-24 武汉大学 Depth significance-based stereopicture just noticeable difference (JND) model building method
CN102905130A (en) * 2012-09-29 2013-01-30 浙江大学 Multi-resolution JND model construction method based on visual perception

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100750138B1 (en) * 2005-11-16 2007-08-21 삼성전자주식회사 Method and apparatus for encoding and decoding video using human visual characteristics

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101621708A (en) * 2009-07-29 2010-01-06 武汉大学 Method for computing perceptible distortion of color image based on DCT field
CN101710995A (en) * 2009-12-10 2010-05-19 武汉大学 Video coding system based on vision characteristic
CN102750706A (en) * 2012-07-13 2012-10-24 武汉大学 Depth significance-based stereopicture just noticeable difference (JND) model building method
CN102905130A (en) * 2012-09-29 2013-01-30 浙江大学 Multi-resolution JND model construction method based on visual perception

Also Published As

Publication number Publication date
CN103475881A (en) 2013-12-25

Similar Documents

Publication Publication Date Title
CN103475881B (en) The image JND threshold value computational methods of view-based access control model attention mechanism in DCT domain
Gu et al. Multiscale natural scene statistical analysis for no-reference quality evaluation of DIBR-synthesized views
Chandler Seven challenges in image quality assessment: past, present, and future research
Beghdadi et al. A survey of perceptual image processing methods
Wan et al. Hybrid JND model-guided watermarking method for screen content images
Wan et al. A novel just noticeable difference model via orientation regularity in DCT domain
He et al. Objective image quality assessment: a survey
Zhang et al. Low-rank decomposition-based restoration of compressed images via adaptive noise estimation
Cedillo-Hernandez et al. A spatiotemporal saliency-modulated JND profile applied to video watermarking
Wang et al. Quaternion representation based visual saliency for stereoscopic image quality assessment
CN103607589B (en) JND threshold value computational methods based on hierarchy selection visual attention mechanism
Beghdadi et al. A critical analysis on perceptual contrast and its use in visual information analysis and processing
Niu et al. Visual saliency’s modulatory effect on just noticeable distortion profile and its application in image watermarking
US20220103869A1 (en) Techniques for limiting the influence of image enhancement operations on perceptual video quality estimations
He et al. Video quality assessment by compact representation of energy in 3D-DCT domain
Yuan et al. Low bit-rate compression of underwater image based on human visual system
CN111105357A (en) Distortion removing method and device for distorted image and electronic equipment
US9230161B2 (en) Multiple layer block matching method and system for image denoising
Yuan et al. Object shape approximation and contour adaptive depth image coding for virtual view synthesis
Abd-Elhafiez Image compression algorithm using a fast curvelet transform
Zhang et al. Local binary pattern statistics feature for reduced reference image quality assessment
Ghanem et al. Segmentation-based perceptual image quality assessment (SPIQA)
Xu et al. Color enhancement algorithm for visual communication posters based on homomorphic filtering
Chetouani et al. Deblocking method using a percpetual recursive filter
Dahiwal et al. An analytical survey on image compression

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20161123

Termination date: 20190912