CN106650737B - Image automatic cropping method - Google Patents
Image automatic cropping method Download PDFInfo
- Publication number
- CN106650737B CN106650737B CN201611041091.9A CN201611041091A CN106650737B CN 106650737 B CN106650737 B CN 106650737B CN 201611041091 A CN201611041091 A CN 201611041091A CN 106650737 B CN106650737 B CN 106650737B
- Authority
- CN
- China
- Prior art keywords
- image
- aesthetic
- map
- candidate
- cropped
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明涉及模式识别、机器学习及计算机视觉技术领域,特别涉及一种图像自动裁剪方法。The invention relates to the technical fields of pattern recognition, machine learning and computer vision, in particular to an image automatic cropping method.
背景技术Background technique
随着计算机技术和数字媒体技术的快速发展,人们对计算机视觉、人工智能、机器感知等领域的需求与期盼也越来越高。图像的自动裁剪作为图像自动编辑中的一项非常重要和常见的任务也得到越来越多的关注和发展。图像自动裁剪技术就是希望能够去除多余的区域,强调感兴趣区域,从而提高图像的整体构图和美感质量。一种有效并且自动的图像裁剪方法不仅能够使人类从繁琐的工作中解放出来,而且还能给一些非专业人士提供一些专业的图像编辑的建议。With the rapid development of computer technology and digital media technology, people's demands and expectations for computer vision, artificial intelligence, machine perception and other fields are also getting higher and higher. As a very important and common task in automatic image editing, automatic image cropping has also received more and more attention and development. The automatic image cropping technology hopes to remove redundant areas and emphasize areas of interest, thereby improving the overall composition and aesthetic quality of the image. An effective and automatic image cropping method can not only free humans from tedious work, but also provide some professional image editing suggestions to some non-professionals.
由于图像裁剪是一项非常主观性的任务,现有的规则很难考虑所有影响因素。传统的图像自动裁剪区域通常使用显著性图来识别图像中的主要区域或感兴趣区域,同时通过制定的一些规则来计算能量函数最小化或学习分类器来寻找裁剪区域。但是这些制定的规则对图像裁剪这一主观性的任务并不够全面,精度也很难达到用户需求。Since image cropping is a very subjective task, it is difficult for existing rules to take into account all influencing factors. Traditional automatic cropping regions of images usually use saliency maps to identify the main regions or regions of interest in the image, and at the same time, some rules are formulated to minimize energy functions or learn classifiers to find cropped regions. However, these rules are not comprehensive enough for the subjective task of image cropping, and the accuracy is difficult to meet user requirements.
有鉴于此,特提出本发明。In view of this, the present invention is proposed.
发明内容SUMMARY OF THE INVENTION
为了解决现有技术中的上述问题,即为了解决如何提高图像自动裁剪的鲁棒性和精度的技术问题而提供一种图像自动裁剪方法。In order to solve the above problems in the prior art, that is, to solve the technical problem of how to improve the robustness and accuracy of automatic image cropping, an automatic image cropping method is provided.
为了实现上述目的,提供了以下技术方案:In order to achieve the above purpose, the following technical solutions are provided:
一种图像自动裁剪方法,所述方法包括:An image automatic cropping method, the method includes:
提取待裁剪图像的美感响应图和梯度能量图;Extract the aesthetic response map and gradient energy map of the image to be cropped;
对所述待裁剪图像密集提取候选裁剪图像;Densely extract candidate cropped images from the to-be-cropped images;
基于所述美感响应图,筛选所述候选裁剪图像;based on the aesthetic response map, screening the candidate cropped images;
基于所述美感响应图和所述梯度能量图,估计筛选出的候选裁剪图像的构图分数,并将得分最高的候选裁剪图像确定为裁剪图像。Based on the aesthetic response map and the gradient energy map, the composition scores of the filtered candidate cropped images are estimated, and the candidate cropped image with the highest score is determined as the cropped image.
进一步地,所述提取待裁剪图像的美感响应图和梯度能量图,具体包括:Further, the extraction of the aesthetic response map and the gradient energy map of the image to be cropped specifically includes:
利用深度卷积神经网络和类别响应映射方法,并采用如下公式提取所述待裁剪图像的所述美感响应图:Using a deep convolutional neural network and a class response mapping method, the following formula is used to extract the aesthetic response map of the to-be-cropped image:
其中,所述M(x,y)表示在空间位置(x,y)处的美感响应值;所述K表示深度卷积神经网络的最后一层卷积层的特征图的总通道个数;所述k表示第k个通道;所述fk(x,y)表示所述第k个通道在所述空间位置(x,y)处的特征值;所述wk表示所述第k个通道的特征图池化后的结果到高美感类别的权值;Wherein, the M(x, y) represents the aesthetic response value at the spatial position (x, y); the K represents the total number of channels of the feature map of the last convolutional layer of the deep convolutional neural network; The k represents the k th channel; the f k (x, y) represents the eigenvalue of the k th channel at the spatial position (x, y); the w k represents the k th channel The result of pooling the feature map of the channel to the weight of the high aesthetic category;
对所述待裁剪图像进行平滑处理,并计算每个像素点的梯度值,从而得到所述梯度能量图。The to-be-cropped image is smoothed, and the gradient value of each pixel is calculated to obtain the gradient energy map.
进一步地,所述深度卷积神经网络通过以下方式训练得到:Further, the deep convolutional neural network is obtained by training in the following manner:
在所述深度卷积神经网络结构的底层设置卷积层;A convolutional layer is arranged at the bottom layer of the deep convolutional neural network structure;
在所述深度卷积神经网络结构的最后一个卷积层之后通过全局平均池化的方法,将每一特征图池化为一个点;After the last convolutional layer of the deep convolutional neural network structure, each feature map is pooled into a point by the method of global average pooling;
连接与美感质量分类类别数相同的全连接层和损失函数。Connect the same number of fully-connected layers and loss functions as the aesthetic quality classification classes.
进一步地,所述基于所述美感响应图,筛选所述候选裁剪图像,具体包括:Further, the screening of the candidate cropped images based on the aesthetic response map specifically includes:
通过如下公式计算所述候选裁剪图像的美感保留分数:The aesthetic preservation score of the candidate cropped image is calculated by the following formula:
其中,所述Sa(C)表示所述候选裁剪图像的所述美感保留分数;所述C表示所述候选裁剪图像;所述(i,j)表示像素的位置;所述I表示原始图像;所述A(i,j)表示在(i,j)位置处的美感响应值;Wherein, the S a (C) represents the aesthetic preservation score of the candidate cropped image; the C represents the candidate cropped image; the (i, j) represents the position of the pixel; the I represents the original image ; The A (i, j) represents the aesthetic response value at the (i, j) position;
将所有候选裁剪图像按照所述美感保留分数从大到小进行排序;Sort all candidate cropped images according to the beauty preservation score from large to small;
选取得分最高的一部分候选裁剪图像。Pick a subset of candidate cropped images with the highest scores.
进一步地,所述基于所述美感响应图和所述梯度能量图,估计筛选出的候选裁剪图像的构图分数,并将得分最高的候选裁剪图像确定为裁剪图像,具体包括:Further, the composition score of the screened candidate cropped images is estimated based on the aesthetic response map and the gradient energy map, and the candidate cropped image with the highest score is determined as the cropped image, specifically including:
基于所述美感响应图和所述梯度能量图建立构图模型;establishing a composition model based on the aesthetic response map and the gradient energy map;
利用所述构图模型估计所述筛选出的候选裁剪图像的构图分数,并将所述得分最高的候选裁剪图像确定为所述裁剪图像。The composition score of the screened candidate cropped images is estimated by using the composition model, and the cropped image candidate with the highest score is determined as the cropped image.
进一步地,所述构图模型通过以下方式获得:Further, the composition model is obtained by:
基于所述美感响应图和所述梯度能量图建立训练图像集;establishing a training image set based on the aesthetic response map and the gradient energy map;
对训练图像进行美感质量类别的标注;Label the training images with aesthetic quality categories;
利用标注的训练图像训练深度卷积神经网络;Train a deep convolutional neural network with annotated training images;
针对所述已标注的训练图像,利用训练好的深度卷积神经网络,提取所述美感响应图和所述梯度能量图的空间金字塔特征;For the marked training image, use the trained deep convolutional neural network to extract the spatial pyramid feature of the aesthetic response map and the gradient energy map;
将提取的空间金字塔特征拼接在一起;stitch together the extracted spatial pyramid features;
利用分类器进行训练,自动学习构图规则,得到构图模型。The classifier is used for training, and the composition rules are automatically learned to obtain the composition model.
本发明实施例提供一种图像自动裁剪方法。该方法包括:提取待裁剪图像的美感响应图和梯度能量图;对待裁剪图像密集提取候选裁剪图像;基于美感响应图,筛选候选裁剪图像;基于美感响应图和梯度能量图,估计筛选出的候选裁剪图像的构图分数,并将得分最高的候选裁剪图像确定为裁剪图像。本方案利用美感响应图去探究图片的美感影响区域,利用美感响应图确定美感保留部分,从而更加最大程度地保留了裁剪图像的高美感质量,同时本方案还利用梯度能量图去分析梯度分布规则,并且基于美感响应图和梯度能量图来评估裁剪图的构图分数。本发明实施例弥补了图像构图表达的缺陷,解决了如何提高图像自动裁剪的鲁棒性和精度的技术问题。本发明实施例能应用于涉及图像自动裁剪的众多领域,包括图像编辑、摄影学及图像重定位等。Embodiments of the present invention provide an automatic image cropping method. The method includes: extracting an aesthetic response map and a gradient energy map of an image to be cropped; intensively extracting candidate crop images from the to-be-cropped image; screening candidate crop images based on the aesthetic response map; estimating the selected candidate images based on the aesthetic response map and the gradient energy map The composition score of the cropped image, and the candidate cropped image with the highest score is determined as the cropped image. This scheme uses the aesthetic response map to explore the aesthetic influence area of the picture, and uses the aesthetic response map to determine the aesthetic retention part, thereby retaining the high aesthetic quality of the cropped image to the greatest extent. At the same time, this scheme also uses the gradient energy map to analyze the gradient distribution rules. , and the composition score of the crop is evaluated based on the aesthetic response map and the gradient energy map. The embodiments of the present invention make up for the defects of image composition expression, and solve the technical problem of how to improve the robustness and accuracy of automatic image cropping. The embodiments of the present invention can be applied to many fields involving automatic image cropping, including image editing, photography, and image relocation.
附图说明Description of drawings
图1是根据本发明实施例的图像自动裁剪方法的流程示意图;1 is a schematic flowchart of an image automatic cropping method according to an embodiment of the present invention;
图2是根据本发明实施例的深度卷积神经网络的结构示意图;2 is a schematic structural diagram of a deep convolutional neural network according to an embodiment of the present invention;
图3a是根据本发明实施例的待裁剪图像示意图;3a is a schematic diagram of an image to be cropped according to an embodiment of the present invention;
图3b是根据本发明实施例的裁剪后的图像示意图。FIG. 3b is a schematic diagram of a cropped image according to an embodiment of the present invention.
具体实施方式Detailed ways
下面结合附图以及具体实施例对本发明实施例解决的技术问题、所采用的技术方案以及实现的技术效果进行清楚、完整的描述。显然,所描述的实施例仅仅是本申请的一部分实施例,并不是全部实施例。基于本申请中的实施例,本领域普通技术人员在不付出创造性劳动的前提下,所获的所有其它等同或明显变型的实施例均落在本发明的保护范围内。本发明实施例可以按照权利要求中限定和涵盖的多种不同方式来具体化。The technical problems solved by the embodiments of the present invention, the technical solutions adopted, and the technical effects achieved will be described clearly and completely below with reference to the accompanying drawings and specific embodiments. Obviously, the described embodiments are only a part of the embodiments of the present application, not all of the embodiments. Based on the embodiments in the present application, all other equivalent or obviously modified embodiments obtained by those of ordinary skill in the art without creative efforts fall within the protection scope of the present invention. Embodiments of the invention can be embodied in a number of different ways as defined and covered by the claims.
深度学习在各个领域得到了快速的发展及很好的效果。本发明实施例考虑利用深度学习去自动学习对图像裁剪重要的影响区域,以自动全面地学习规则,从而使得在裁剪时尽可能地保留高美感区域。Deep learning has achieved rapid development and good results in various fields. The embodiment of the present invention considers using deep learning to automatically learn an important influence area for image cropping, so as to automatically and comprehensively learn the rules, so that the high aesthetic area is preserved as much as possible during cropping.
为此,本发明实施例提供一种自动图像裁剪方法。图1示例性地示出了图像自动裁剪方法的流程。如图1所示,该方法可以包括:To this end, an embodiment of the present invention provides an automatic image cropping method. FIG. 1 exemplarily shows the flow of an image automatic cropping method. As shown in Figure 1, the method may include:
S100:提取待裁剪图像的美感响应图和梯度能量图。S100: Extract the aesthetic response map and the gradient energy map of the image to be cropped.
具体地,本步骤可以包括:Specifically, this step may include:
S101:利用深度卷积神经网络和类别响应映射方法,并采用如下公式提取待裁剪图像的美感响应图:S101: Using a deep convolutional neural network and a class response mapping method, and using the following formula to extract the aesthetic response map of the image to be cropped:
其中,M(x,y)表示在空间位置(x,y)处的美感响应值;K表示训练好的深度卷积神经网络的最后一层卷积层的特征图f的总通道个数;k表示第k个通道;fk(x,y)表示第k个通道在空间位置(x,y)处的特征值;wk表示第k个通道的特征图池化后的结果到高美感类别的权值。Among them, M(x,y) represents the aesthetic response value at the spatial position (x,y); K represents the total number of channels of the feature map f of the last convolutional layer of the trained deep convolutional neural network; k represents the k-th channel; f k (x, y) represents the feature value of the k-th channel at the spatial position (x, y); w k represents the result of pooling the feature map of the k-th channel to high aesthetics category weights.
上述步骤在提取美感响应图时可以根据实际需要训练深度卷积神经网络。深度卷积神经网络的训练可以通过以下方式进行:The above steps can train a deep convolutional neural network according to actual needs when extracting the aesthetic response map. Training of deep convolutional neural networks can be done in the following ways:
步骤1:在深度卷积神经网络结构的底层设置卷积层。Step 1: Set up convolutional layers at the bottom layer of the deep convolutional neural network structure.
步骤2:在深度卷积神经网络结构的最后一个卷积层之后通过全局平均池化的方法,将每一个特征图池化为一个点。Step 2: After the last convolutional layer of the deep convolutional neural network structure, each feature map is pooled into a point by the method of global average pooling.
步骤3:连接一个与美感质量分类类别数相同的全连接层和损失函数。Step 3: Connect a fully connected layer and loss function with the same number of categories as the aesthetic quality classification.
图2示例性地示出了一个深度卷积神经网络结构。Figure 2 exemplarily shows a deep convolutional neural network structure.
通过步骤1-3可以训练一个在美感质量分类任务下的深度卷积神经网络模型。然后,利用为美感质量分类任务训练好的深度卷积神经网络和类别响应映射方法;再采用上述公式,可以计算在高美感类别下待裁剪图像的美感响应图M。Through steps 1-3, a deep convolutional neural network model under the task of aesthetic quality classification can be trained. Then, using the deep convolutional neural network trained for the aesthetic quality classification task and the category response mapping method; and then using the above formula, the aesthetic response map M of the image to be cropped under the high aesthetic category can be calculated.
S102:对待裁剪图像进行平滑处理,并计算每个像素点的梯度值,从而得到梯度能量图。S102: Smooth the to-be-cropped image, and calculate the gradient value of each pixel to obtain a gradient energy map.
S110:对待裁剪图像密集提取候选裁剪图像。S110: Densely extract candidate cropped images from the to-be-cropped images.
这里,可以采用小于图像大小的所有大小的滑动窗口,对待裁剪图像密集提取候选裁剪窗口,通过候选裁剪窗口提取出候选裁剪图像。Here, sliding windows of all sizes smaller than the image size can be used, and candidate cropping windows are intensively extracted from the to-be-cropped image, and candidate cropping images are extracted through the candidate cropping windows.
S120:基于美感响应图,筛选候选裁剪图像。S120: Screen candidate cropped images based on the aesthetic response map.
具体地,本步骤可以包括:Specifically, this step may include:
S121:通过如下公式计算候选裁剪图像的美感保留分数:S121: Calculate the beauty preservation score of the candidate cropped image by the following formula:
其中,Sa(C)表示候选裁剪图像的美感保留分数;C表示候选裁剪图像;(i,j)表示像素的位置;I表示原始图像;A(i,j)表示在(i,j)处的美感响应值。Among them, S a (C) represents the aesthetic preservation score of the candidate cropped image; C represents the candidate cropped image; (i, j) represents the position of the pixel; I represents the original image; A (i, j) represents at (i, j) Aesthetic response value at .
通过本步骤可以构建美感保留模型。将候选裁剪窗口经过美感保留模型筛选出美感保留分数较高的候选窗口。Through this step, an aesthetic preservation model can be constructed. The candidate cropping window is filtered through the aesthetics preservation model to select the candidate window with higher aesthetics preservation score.
S122:将所有候选裁剪图像按照美感保留分数从大到小进行排序。S122: Sort all candidate cropped images in descending order according to their beauty preservation scores.
S123:选取得分最高的一部分候选裁剪图像。S123: Select a part of candidate cropped images with the highest score.
例如:实际应用中可以设置保留前10000个候选裁剪窗口中的候选裁剪图像。For example, in practical applications, the candidate cropped images in the first 10,000 cropped candidate windows can be set to be retained.
S130:基于美感响应图和梯度能量图,估计筛选出的候选裁剪图像的构图分数,并将得分最高的候选裁剪图像确定为裁剪图像。S130: Based on the aesthetic response map and the gradient energy map, estimate the composition score of the selected candidate cropped images, and determine the candidate cropped image with the highest score as the cropped image.
具体地,本步骤可以通过步骤S131至步骤S133来实现。Specifically, this step can be implemented through steps S131 to S133.
S131:基于美感响应图和梯度能量图建立构图模型。S131: Establish a composition model based on the aesthetic response map and the gradient energy map.
本步骤在建立构图模型时可以根据实际情况训练构图模型。在训练构图模型的过程中,训练数据可以采用构图较好的图像作为正样本,而将有构图缺陷的图像作为负样本。In this step, the composition model can be trained according to the actual situation when the composition model is established. In the process of training the composition model, the training data can use images with good composition as positive samples, and images with composition defects as negative samples.
可以通过以下方式来训练构图模型:The composition model can be trained in the following ways:
步骤a:基于美感响应图和梯度能量图建立训练图像集。Step a: Build a training image set based on the aesthetic response map and the gradient energy map.
步骤b:对训练图像进行美感质量类别的标注。Step b: Label the training images with aesthetic quality categories.
步骤c:利用标注的训练图像训练深度卷积神经网络。Step c: Train a deep convolutional neural network with annotated training images.
本步骤的训练过程可以参考上述步骤1至步骤3,在此不再赘述。For the training process of this step, reference may be made to the above-mentioned steps 1 to 3, which will not be repeated here.
步骤d:针对已标注的训练图像,利用训练好的深度卷积神经网络,提取美感响应图和梯度能量图的空间金字塔特征。Step d: For the labeled training images, use the trained deep convolutional neural network to extract the spatial pyramid features of the aesthetic response map and the gradient energy map.
步骤e:将提取的空间金字塔特征拼接在一起。Step e: stitch together the extracted spatial pyramid features.
步骤f:利用分类器进行训练,自动学习构图规则,得到构图模型。Step f: Use the classifier for training, automatically learn composition rules, and obtain a composition model.
其中,分类器例如可以采用支持向量机分类器。The classifier can be, for example, a support vector machine classifier.
S132:利用构图模型估计筛选出的候选裁剪图像的构图分数,并将得分最高的候选裁剪图像确定为裁剪图像。S132: Use the composition model to estimate the composition score of the selected candidate cropped images, and determine the candidate cropped image with the highest score as the cropped image.
图3a示例性地示出了待裁剪图像;图3b示例性地示出了裁剪后的图像。Fig. 3a exemplarily shows the image to be cropped; Fig. 3b exemplarily shows the cropped image.
下面再以一优选实施例来更好地说明本发明。Hereinafter, the present invention will be better described with a preferred embodiment.
步骤A:将标注有美感质量类别的图像数据集送入深度卷积神经网络进行美感质量类别模型训练。Step A: Send the image dataset marked with the aesthetic quality category into the deep convolutional neural network for training the aesthetic quality category model.
步骤B:将标注有构图类别的图像数据集输入训练好的深度卷积神经网络,提取最后一层卷积层的特征图,并计算美感响应图,同时计算美感梯度图,然后采用支持向量机分类器训练构图模型。Step B: Input the image dataset marked with the composition category into the trained deep convolutional neural network, extract the feature map of the last convolutional layer, calculate the aesthetic response map, and calculate the aesthetic gradient map at the same time, and then use the support vector machine. The classifier trains the composition model.
步骤C:对待测试图像提取美感响应图和梯度能量图。Step C: Extract the aesthetic response map and the gradient energy map from the test image.
本步骤的提取方法可参考训练阶段的方法。For the extraction method in this step, refer to the method in the training phase.
步骤D:密集采集待测试图像的候选裁剪窗口。Step D: Densely collect candidate cropping windows of the image to be tested.
举例来说,在1000×1000的待测试图像上,利用间隔为30个像素的滑动窗口进行采集或提取。For example, on a 1000×1000 image to be tested, a sliding window with an interval of 30 pixels is used to collect or extract.
步骤E:利用美感保留模型筛选候选裁剪窗口。Step E: Screen candidate cropping windows using an aesthetic preservation model.
本步骤利用美感保留模型计算密集采集到的候选裁剪窗口的美感保留分数,筛选出美感分类最高的一部分候选裁剪窗口,例如:筛选出10000个候选裁剪窗口。In this step, the beauty retention model is used to calculate the beauty retention scores of the densely collected candidate cropping windows, and a part of the candidate cropping windows with the highest aesthetics classification is selected, for example, 10,000 candidate cropping windows are screened out.
步骤F:利用构图模型评估筛选出的候选裁剪窗口。Step F: Use the composition model to evaluate the screened candidate cropping windows.
本步骤采集训练阶段训练好的构图模型去评估筛选出的候选裁剪窗口的构图分数,将得分最高的作为最后的裁剪窗口,从而得到裁剪图像。In this step, the composition model trained in the training stage is collected to evaluate the composition scores of the selected candidate cropping windows, and the cropped image is obtained by taking the one with the highest score as the final cropping window.
综上所述,本发明实施例提供的方法很好地利用了美感响应图和梯度能量图来最大程度地保留美感质量和图像的构图规则,得到更加鲁棒,精度更高的图像的自动裁剪性能,进而说明了美感响应图和梯度能量图对于图像自动裁剪的有效性。To sum up, the method provided by the embodiment of the present invention makes good use of the aesthetic response map and the gradient energy map to maximize the preservation of the aesthetic quality and the composition rules of the image, so as to obtain a more robust and accurate automatic cropping of the image. performance, which further illustrates the effectiveness of aesthetic response maps and gradient energy maps for automatic image cropping.
上述实施例中虽然按照上述先后次序描述了本发明实施例提供的方法,但是本领域技术人员可以理解,为了实现本实施例的效果,还可以以诸如并行或颠倒次序等不同的顺序来执行,这些简单的变化都在本发明的保护范围之内。Although the method provided by the embodiment of the present invention is described in the above-mentioned order in the above-mentioned embodiment, those skilled in the art can understand that, in order to achieve the effect of this embodiment, it can also be executed in different orders such as parallel or reversed order, etc. These simple variations are within the scope of the present invention.
以上所述,仅为本发明中的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉该技术的人在本发明所揭露的技术范围内,可理解想到的变换或替换,都应涵盖在本发明的包含范围之内,因此,本发明的保护范围应该以权利要求书的保护范围为准。The above is only a specific embodiment of the present invention, but the protection scope of the present invention is not limited to this, any person familiar with the technology can understand the transformation or replacement that comes to mind within the technical scope disclosed by the present invention, All should be included within the scope of the present invention, therefore, the protection scope of the present invention should be subject to the protection scope of the claims.
Claims (5)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201611041091.9A CN106650737B (en) | 2016-11-21 | 2016-11-21 | Image automatic cropping method |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201611041091.9A CN106650737B (en) | 2016-11-21 | 2016-11-21 | Image automatic cropping method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN106650737A CN106650737A (en) | 2017-05-10 |
| CN106650737B true CN106650737B (en) | 2020-02-28 |
Family
ID=58811471
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201611041091.9A Active CN106650737B (en) | 2016-11-21 | 2016-11-21 | Image automatic cropping method |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN106650737B (en) |
Families Citing this family (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107317962B (en) * | 2017-05-12 | 2019-11-08 | 广东网金控股股份有限公司 | An intelligent photo cropping and composition system and its application method |
| CN107392244B (en) * | 2017-07-18 | 2020-08-28 | 厦门大学 | Image aesthetic feeling enhancement method based on deep neural network and cascade regression |
| CN107545576A (en) * | 2017-07-31 | 2018-01-05 | 华南农业大学 | Image edit method based on composition rule |
| CN108154464B (en) * | 2017-12-06 | 2020-09-22 | 中国科学院自动化研究所 | Method and device for automatically clipping picture based on reinforcement learning |
| CN108566512A (en) * | 2018-03-21 | 2018-09-21 | 珠海市魅族科技有限公司 | A kind of intelligence image pickup method, device, computer equipment and readable storage medium storing program for executing |
| CN109523503A (en) * | 2018-09-11 | 2019-03-26 | 北京三快在线科技有限公司 | A kind of method and apparatus of image cropping |
| CN109518446B (en) * | 2018-12-21 | 2021-01-01 | 季华实验室 | Intelligent cutting method of cutting machine |
| CN109886317B (en) * | 2019-01-29 | 2021-04-27 | 中国科学院自动化研究所 | A general image aesthetic evaluation method, system and device based on attention mechanism |
| WO2020186385A1 (en) * | 2019-03-15 | 2020-09-24 | 深圳市大疆创新科技有限公司 | Image processing method, electronic device, and computer-readable storage medium |
| CN110062173A (en) * | 2019-03-15 | 2019-07-26 | 北京旷视科技有限公司 | Image processor and image processing method, equipment, storage medium and intelligent terminal |
| CN111684488A (en) * | 2019-05-22 | 2020-09-18 | 深圳市大疆创新科技有限公司 | Image cropping method, device and photographing device |
| CN112839167B (en) * | 2020-12-30 | 2023-06-30 | Oppo(重庆)智能科技有限公司 | Image processing method, device, electronic equipment and computer readable medium |
| CN113436224B (en) * | 2021-06-11 | 2022-04-26 | 华中科技大学 | Intelligent image clipping method and device based on explicit composition rule modeling |
| CN114119373A (en) * | 2021-11-29 | 2022-03-01 | 广东维沃软件技术有限公司 | Image cropping method and device and electronic equipment |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104717413A (en) * | 2013-12-12 | 2015-06-17 | 北京三星通信技术研究有限公司 | Shooting assistance method and equipment |
| CN105488758A (en) * | 2015-11-30 | 2016-04-13 | 河北工业大学 | Image scaling method based on content awareness |
| CN105528786A (en) * | 2015-12-04 | 2016-04-27 | 小米科技有限责任公司 | Image processing method and device |
| CN105787966A (en) * | 2016-03-21 | 2016-07-20 | 复旦大学 | An aesthetic evaluation method for computer pictures |
| CN105894025A (en) * | 2016-03-30 | 2016-08-24 | 中国科学院自动化研究所 | Natural image aesthetic feeling quality assessment method based on multitask deep learning |
-
2016
- 2016-11-21 CN CN201611041091.9A patent/CN106650737B/en active Active
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104717413A (en) * | 2013-12-12 | 2015-06-17 | 北京三星通信技术研究有限公司 | Shooting assistance method and equipment |
| CN105488758A (en) * | 2015-11-30 | 2016-04-13 | 河北工业大学 | Image scaling method based on content awareness |
| CN105528786A (en) * | 2015-12-04 | 2016-04-27 | 小米科技有限责任公司 | Image processing method and device |
| CN105787966A (en) * | 2016-03-21 | 2016-07-20 | 复旦大学 | An aesthetic evaluation method for computer pictures |
| CN105894025A (en) * | 2016-03-30 | 2016-08-24 | 中国科学院自动化研究所 | Natural image aesthetic feeling quality assessment method based on multitask deep learning |
Non-Patent Citations (2)
| Title |
|---|
| 基于并行深度卷积神经网络的图像美感分类;王伟凝 等;《自动化学报》;20160630;第42卷(第6期);第904-913页 * |
| 相片中重要对象布局优化系统;侯丹红;《中国优秀硕士学位论文全文数据库 信息科技辑》;20110615(第6期);第17-28页 * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN106650737A (en) | 2017-05-10 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN106650737B (en) | Image automatic cropping method | |
| CN107665492B (en) | A deep network-based tissue segmentation method for colorectal panoramic digital pathological images | |
| WO2018090355A1 (en) | Method for auto-cropping of images | |
| CN106960195B (en) | Crowd counting method and device based on deep learning | |
| Friedland et al. | SIOX: Simple interactive object extraction in still images | |
| CN109801256A (en) | A kind of image aesthetic quality appraisal procedure based on area-of-interest and global characteristics | |
| WO2020007307A1 (en) | Sky filter method for panoramic images and portable terminal | |
| CN111027547A (en) | An automatic detection method for multi-scale and polymorphic objects in two-dimensional images | |
| CN110570435B (en) | Method and device for carrying out damage segmentation on vehicle damage image | |
| CN108960404B (en) | Image-based crowd counting method and device | |
| CN110569747A (en) | A method to quickly count the number of rice ears in field rice using image pyramid and Faster-RCNN | |
| CN111382766A (en) | A device fault detection method based on Faster R-CNN | |
| WO2018076212A1 (en) | De-convolutional neural network-based scene semantic segmentation method | |
| CN109767422A (en) | Pipe detection recognition methods, storage medium and robot based on deep learning | |
| CN109284733A (en) | A method for monitoring negative behavior of shopping guide based on yolo and multi-task convolutional neural network | |
| CN107066916A (en) | Scene Semantics dividing method based on deconvolution neutral net | |
| CN111222519A (en) | Model construction, method and device for extracting line draft of layered painted cultural relics | |
| CN116596877B (en) | A Bone Age Intelligent Assessment Method Based on Attention Mechanism | |
| JP7350208B2 (en) | Image processing device, image processing method, and program | |
| CN109741351A (en) | A class-sensitive edge detection method based on deep learning | |
| CN110796716A (en) | Image coloring method based on multiple residual error networks and regularized transfer learning | |
| CN107315999A (en) | A kind of tobacco plant recognition methods based on depth convolutional neural networks | |
| CN111540203B (en) | Method for adjusting green light passing time based on fast-RCNN | |
| CN109460767A (en) | Rule-based convex print bank card number segmentation and recognition methods | |
| CN109978858B (en) | Double-frame thumbnail image quality evaluation method based on foreground detection |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |



