CN110519509A - Composition evaluation method, method for imaging, device, electronic equipment, storage medium - Google Patents
Composition evaluation method, method for imaging, device, electronic equipment, storage medium Download PDFInfo
- Publication number
- CN110519509A CN110519509A CN201910708006.7A CN201910708006A CN110519509A CN 110519509 A CN110519509 A CN 110519509A CN 201910708006 A CN201910708006 A CN 201910708006A CN 110519509 A CN110519509 A CN 110519509A
- Authority
- CN
- China
- Prior art keywords
- composition
- composition frame
- frame
- assessment
- evaluation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
- H04N23/632—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Biophysics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Signal Processing (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Image Analysis (AREA)
Abstract
The present invention provides composition evaluation method, method for imaging, device, electronic equipment, storage mediums, comprising: extracts the characteristic information of image in the frame of composition frame, the composition frame includes photographic subjects;The composition frame is evaluated based on the characteristic information, obtains the first assessment of the composition frame;The composition frame is evaluated based on the composition evaluation model of deep learning method, obtains the second assessment of the composition frame;According to the first assessment and the second assessment, the overall merit of the composition frame is obtained.The present invention has done comprehensive evaluation to photography composition from the shallow-layer feature and further feature of image, so that output meets the photo of human aesthetic's requirement.
Description
Technical field
The present invention relates to automatic photography technical field, espespecially composition evaluation method, method for imaging, device, electronic equipment, deposit
Storage media.
Background technique
In production and living practice, people using the cameras such as camera by optical imaging concept, with obtain image,
Image, and record image, image.In general, needing such as photographer by manual operation camera, to obtain default
Image, the image of photographic subjects.Typically, the basis of photography is composition, that is to say selection photographic subjects and will shoot mesh
It is marked with a certain size, ratio is placed in corresponding composition frame, and finally showed in picture, and then obtain to hold accordingly
In the material for carrying photographic subjects relevant information.
Good composition can make photo clearly express artistic conception and site environment atmosphere, however how evaluate the quality of composition
It is a very difficult and complicated thing, composition evaluation is a very subjective problem in itself.Traditional composition evaluation side
Method is mostly the experience in conjunction with photographer, does aesthetic assessment to composition from bottom visual information.How from the angle of quantization more comprehensively
The quality of composition is evaluated on ground, and being one is worth the problem of further exploring.
Summary of the invention
An object of the present invention is to provide composition evaluation side to overcome at least partly deficiency existing in the prior art
Method, method for imaging, device, electronic equipment, storage medium do comprehensively photography composition from the shallow-layer feature and further feature of image
Evaluation, thus output meet human aesthetic requirement photo.
Technical solution provided by the invention is as follows:
A kind of photography composition evaluation method, comprising: extract the characteristic information of image in the frame of composition frame, the composition frame packet
Contain photographic subjects;The composition frame is evaluated based on the characteristic information, obtains the first assessment of the composition frame;
The composition frame is evaluated based on the composition evaluation model of deep learning method, obtains commenting for the second time for the composition frame
Valence;According to the first assessment and the second assessment, the overall merit of the composition frame is obtained.
It is further preferred that described carry out evaluating it based on the composition evaluation model of deep learning method to the composition frame
Before include: the label for obtaining several images for being added to composition frame and each composition frame, obtain pattern data collection;With described
The further feature of pattern data collection training deep learning network, the deep learning network is evaluated for composition;When the depth
When learning network restrains, composition evaluation model is obtained.
It is further preferred that the further feature of the deep learning network is evaluated for composition further include: described
The further feature of deep learning network and at least one shallow-layer feature are combined to be evaluated for composition.
It is further preferred that described be based on the characteristic information to carry out evaluation to the composition frame including: based on the spy
The random forest disaggregated model of reference breath evaluates the composition frame.
It is further preferred that the characteristic information of image includes: according in photographic subjects in the frame for extracting composition frame
The range information in four crosspoints of nine grids, obtains composition of geometry characterization factor in the heart and composition frame;And/or when shooting mesh
When being designated as one, the vision equilibrium characteristic factor is obtained according to the facial orientation of the photographic subjects;When photographic subjects are more people,
The vision equilibrium characteristic factor is obtained according to personage's center of gravity of the photographic subjects;And/or when photographic subjects are personage, according to
Integrity degree of the human body key node of the photographic subjects in composition frame, obtains the integrity feature factor of human body key point;
And/or when photographic subjects are personage, according to the expression of the photographic subjects, obtain the expressive features factor;And/or work as shooting
When target is personage, according to the statistics with histogram of the background pixel of composition frame, background color terseness characterization factor is obtained.
The present invention also provides a kind of photography composition evaluation methods, comprising: the characteristic information of image in the frame of composition frame is extracted,
The composition frame includes photographic subjects;The composition frame is evaluated based on the characteristic information, obtains the composition frame
The first assessment;Judge whether the first assessment meets preset condition;When the first assessment is unsatisfactory for default item
When part, the overall merit of the composition frame is obtained according to the first assessment;When the first assessment meets preset condition
When, the composition frame is evaluated using the composition evaluation model based on deep learning method, obtains the of the composition frame
Second evaluation;And according to the first assessment and the second assessment, obtain the overall merit of the composition frame.
The present invention also provides a kind of automatic photography methods, comprising: generation includes the composition frame of photographic subjects;According to aforementioned
Any one of described in photography composition evaluation method the composition frame is evaluated, obtain overall merit;According to the synthesis
Evaluation controls the output of the composition frame;When photographic unit receives the composition frame, photograph according to the composition frame.
The present invention also provides a kind of photography composition evaluating apparatus, comprising: characteristic extracting module, for extracting the frame of composition frame
The characteristic information of interior image, the composition frame includes photographic subjects;First evaluation module, for being based on the characteristic information pair
The composition frame is evaluated, and the first assessment of the composition frame is obtained;Second evaluation module, for being based on deep learning side
The composition evaluation model of method evaluates the composition frame, obtains the second assessment of the composition frame;Overall merit module,
For obtaining the overall merit of the composition frame according to the first assessment and the second assessment.
It is further preferred that further include: model construction module, for obtain several images for being added to composition frame and
The label of each composition frame, obtains pattern data collection;And with the pattern data collection training deep learning network, the depth
The further feature for spending learning network is evaluated for composition;When the deep learning network convergence, composition evaluation model is obtained.
It is further preferred that the model construction module, is further used for the pattern data collection training deep learning
Network, the further feature of the deep learning network and at least one shallow-layer feature are combined to be evaluated for composition.
It is further preferred that first evaluation module, is further used for the random forest based on the characteristic information point
Class model evaluates the composition frame.
It is further preferred that the characteristic extracting module, is further used for according in the center of photographic subjects and composition frame
The range information in four crosspoints of nine grids, obtains composition of geometry characterization factor;And/or when photographic subjects is one, root
The vision equilibrium characteristic factor is obtained according to the facial orientation of the photographic subjects;When photographic subjects are more people, according to the shooting
Personage's center of gravity of target obtains the vision equilibrium characteristic factor;And/or when photographic subjects are personage, according to the photographic subjects
Integrity degree of the human body key node in composition frame, obtain the integrity feature factor of human body key point;And/or work as shooting
When target is personage, according to the expression of the photographic subjects, the expressive features factor is obtained;And/or when photographic subjects are personage
When, according to the statistics with histogram of the background pixel of composition frame, obtain background color terseness characterization factor.
The present invention also provides a kind of photography composition evaluating apparatus, including characteristic extracting module, for extracting the frame of composition frame
The characteristic information of interior image, the composition frame includes photographic subjects;First evaluation module, for being based on the characteristic information pair
The composition frame is evaluated, and the first assessment of the composition frame is obtained;Whether judgment module judges the first assessment
Meet preset condition;Overall merit module, for when the first assessment is unsatisfactory for preset condition, according to the first time
Evaluation obtains the overall merit of the composition frame;Second evaluation module, for when the first assessment meets preset condition,
The composition frame is evaluated using the composition evaluation model based on deep learning method, obtains second of the composition frame
Evaluation;The overall merit module is further used for obtaining the structure according to the first assessment and the second assessment
The overall merit of picture frame.
The present invention also provides a kind of automatic photographing devices, comprising: composition frame generation module includes shooting mesh for generating
Target composition frame;Photography composition evaluating apparatus described in any one of aforementioned obtains comprehensive for evaluating the composition frame
Close evaluation;Composition frame output module, for controlling the output of the composition frame according to the overall merit;Photographic unit is used for
When receiving the composition frame, photograph according to the composition frame.
The present invention also provides a kind of electronic equipment, comprising: memory, for storing computer program;Processor, for transporting
The row computer program, realize it is any one of aforementioned described in photography composition evaluation method.
The present invention also provides a kind of computer readable storage mediums, are stored thereon with computer program, the computer journey
Realized when sequence is executed by processor it is any one of aforementioned described in photography composition evaluation method.
The composition evaluation method that there is provided through the invention, method for imaging, device, electronic equipment, storage medium, can bring
Below the utility model has the advantages that
1, photography composition evaluation method provided by the invention, not only from local message or bottom visual information but also from entirety
Information or profound information evaluate the quality of composition, traditional characteristic information have both been considered, it is contemplated that the sum of picture
Humorous property, the integrality of patterned matter, evaluation is more comprehensively.
2, automatic photography method provided by the invention promotes the promotion of patterning quality by the promotion that composition is evaluated, thus
It avoids exporting low-quality photo, and then low quality photo is avoided to occupy memory source, increase loss etc..
3, the present invention can substitute manpower and realize automatic shooting, avoid Human Physiology tired and subjective factor is to the shadow of shooting
It rings.
Detailed description of the invention
Below by clearly understandable mode, preferred embodiment is described with reference to the drawings, to a kind of composition evaluation method, takes the photograph
Image method, device, electronic equipment, storage medium above-mentioned characteristic, technical characteristic, advantage and its implementation give furtherly
It is bright.
Fig. 1 is a kind of flow chart of one embodiment of photography composition evaluation method of the invention;
Fig. 2 is a kind of flow chart of another embodiment of photography composition evaluation method of the invention;
Fig. 3 is a kind of flow chart of another embodiment of photography composition evaluation method of the invention;
Fig. 4 is a kind of flow chart of one embodiment of automatic photography method of the invention;
Fig. 5 is a kind of structural schematic diagram of one embodiment of photography composition evaluating apparatus of the invention;
Fig. 6 is a kind of structural schematic diagram of another embodiment of photography composition evaluating apparatus of the invention;
Fig. 7 is a kind of structural schematic diagram of another embodiment of photography composition evaluating apparatus of the invention;
Fig. 8 is the structural schematic diagram of one embodiment of a kind of electronic equipment of the invention;
Fig. 9 is an example of visual balance in a kind of photography composition evaluation method of the invention;
Figure 10 is a kind of structural schematic diagram of one embodiment of automatic photographing device of the invention;
Figure 11 is a kind of schematic perspective view of one embodiment of automatic photographing device of the invention.
Drawing reference numeral explanation:
100. photography composition evaluating apparatus, 110. characteristic extracting modules, 120. first evaluation modules, 130. second evaluation moulds
Block, 140. overall merit modules, 150. model construction modules, 160. judgment modules, 200. composition frame generation modules, 300. compositions
Frame output module, 500. automatic photographing devices, 502. photographic units, 400. electronic equipments, 410. memories, 420. processors,
430. computer programs, 440. buses.
Specific embodiment
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, Detailed description of the invention will be compareed below
A specific embodiment of the invention.
It is understood that as needed, photographic subjects can be children or adult, be also possible to it is one or more its
He is personage, can also be other animals, plant, landscape or building etc., can also be any combination in aforementioned object." composition
Frame " refers to the closure frame with certain shapes and size, interior for limiting shape, the size of photographic picture, to realize
Selection to the relative position of photographic subjects.In general, " composition frame " is rectangle frame.More specifically, " composition frame " is left and right
Width is greater than the rectangle frame of upper-lower height.As needed, when such as camera is rotated by 90 ° realization and erects bat, above and below " composition frame "
Height is then greater than left and right width.Correspondingly, when composition frame is selected, and when the images such as output photo, the shape of composition frame
Corresponding to picture.
In addition, unless otherwise instructed, " upper and lower " that occurs in the present invention, " left and right ", " forward and backward ", " one, another " etc.
It is relative concept.In addition, the term " first " occurred in the present invention, " second " etc. are only used for description conveniently, and it is not understood to
Indication or suggestion relative importance clearly limits sequencing.
In one embodiment of the invention, as shown in Figure 1, a kind of photography composition evaluation method, comprising:
Step S200 extracts the characteristic information of image in the frame of composition frame, and the composition frame includes photographic subjects.
Step S300 is based on the characteristic information and evaluates the composition frame, and the first time for obtaining the composition frame comments
Valence.
Specifically, as needed, photographic subjects can be adult or children, it can be other one or more personages.It adopts
It whether can be identified in composition frame with known face recognition technology comprising photographic subjects.It but include the composition frame of photographic subjects
It is not necessarily a good composition, in order to avoid generating low quality photo, output as far as possible meets the photo that human aesthetic requires,
The composition to composition frame is needed to evaluate.
Feature information extraction is carried out to image in the frame of composition frame first.Characteristic information is usually to be passed through according to traditional photography
It tests, is mostly extracted from the bottom visual information of composition, and can be by limited or simple mathematical knowledge expression, for example, shooting
When target is personage, position feature of the photographic subjects in composition frame is (for example, the center and each side of composition frame that pass through photographic subjects
Range information reflection), size characteristic of the photographic subjects in composition frame.Position based on photographic subjects in composition frame is special
Sign, it can be determined that photographic subjects are the edges in center or composition frame, if being in the edge of composition frame, structure
Scheme undesirable;If opposite center, composition meet the aesthetic habit of people in picture;Based on photographic subjects in composition
Size characteristic in frame, it can be determined that whether photographic subjects size is too small in composition frame, if size is too small, composition is paid no attention to
Think.
Features above information is one or more mathematical expressions of parts of traditional photography experience, and this mathematical expression needs people
Work is pre-defined, and machine extracts corresponding characteristic information value according to the mathematical expression predetermined from specific composition frame,
The composition of composition frame is evaluated further according to these characteristic information values, obtains the first assessment.
Step S400 evaluates the composition frame based on the composition evaluation model of deep learning method, obtains the structure
The second assessment of picture frame.
Specifically, manual features extraction comparison is laborious, and the feature that can manually extract is also than relatively limited, mostly shallow-layer
Feature has some features (such as further feature) that can not also express by limited mathematical knowledge, for example image is presented
Artistic conception, concordance of picture etc..
Deep learning is a kind of machine learning algorithm, and benefit is not need manually to extract feature, is learnt by supervised
Feature can be automatically extracted, shallow-layer feature can not only be extracted, moreover it is possible to extract further feature, pass through complicated mathematical knowledge table
Up to out.Deep learning network is a kind of neural network with more hidden layers, is with classical convolutional neural networks Alexnet
Example has 1 input layer, 6 hidden layers, 1 output layer, and totally 8 layers, every layer has thousands of a neurons, and every layer of output is the one of data
Kind feature, and feature hierarchy is improved with the increase of processing level, for example, first layer is the edge feature of image, the second layer is
The color characteristic of image, it is the concordance feature of image that third layer, which is the 8th layer of brightness ... ... of image, network it is former
Layer reflects the local feature of image, is called shallow-layer feature, such as the features such as edge, color, brightness of image, network it is last
One layer, i.e. the top layer global feature that reflects image, are called further feature, such as the concordance feature of image.
When being trained to deep learning network, only further feature can be evaluated applied to composition, it can also be by deep layer
Feature and shallow-layer feature combined application are evaluated in composition.When deep learning network convergence, composition evaluation model is obtained.Use this
Composition evaluation model evaluates composition frame, obtains the second assessment of the composition frame.
Step S500 obtains the overall merit of the composition frame according to the first assessment and the second assessment.
Specifically, the composition based on deep learning evaluates (i.e. the second assessment) cannot replace the first assessment completely, i.e.,
Composition evaluation model is set to consider further feature and shallow-layer feature, also can be problematic, the reason is that: one, deep learning learns out
Feature lack interpretation, it is why good, it is why bad, can not be explicitly stated;Two, the spy that deep learning learns out
Sign depends greatly on data set, and the distribution of data set and abundant degree will limit its effect;Exactly because above-mentioned two
There is uncontrollable situation, i.e. depth in reason, the generalization ability that will lead to the composition evaluation model based on deep learning method
Algorithm is practised to lack in common sense.
The first assessment is evaluated from the feature of Manual definition.The feature of Manual definition, i.e. traditional characteristic, mostly shallowly
Layer feature, interpretation be it is guaranteed, generalization ability is also more secure than deep learning, all for clearly good feature
Bad feature is specified, it is more suitable for going description with traditional characteristic, and can also largely reduce calculation amount.But it passes
System feature also has the problem of itself, is manually also difficult to define further feature, such as the concordance of whole picture etc. at present, and this
A little high-level things, exactly deep learning neural network are relatively good at.
The second assessment considers the further feature that machine deep learning arrives, and the two, which combines, just can more fully be commented
Valence had both considered interpretation feature, it is also considered that not describable further feature (i.e. high-level semantics features), it in this way could be complete
The quality for removing evaluation composition in face, existing details also have entirety, and generalization ability is more secure.
Overall merit is obtained according to the first assessment and the second assessment.It, just can be further only when overall merit is high
Composition frame is exported, is photographed;The promotion of patterning quality is promoted by the promotion that composition is evaluated, to avoid output low quality
Photo, and then avoid low quality photo occupy memory source, increase loss etc..
In another embodiment of the present invention, as shown in Fig. 2, a kind of photography composition evaluation method, comprising:
Step S100 obtains the label of several images for being added to composition frame and each composition frame, obtains pattern data
Collection;
The further feature of the step S110 pattern data collection training deep learning network, the deep learning network is used
It is evaluated in composition;
Step S130 obtains composition evaluation model when the deep learning network convergence.
Specifically, using large-scale dataset, such as millions of images, composition frame is added to every image, and stamp
Label obtains pattern data collection.Since composition evaluation is subjective, there are apparent individual differences, therefore, for composition frame
The mark of interior image is generally labeled same piece image using more people, takes proprietary average mark as image later
Final label.
Deep learning network structure using classical convolutional neural networks, such as Alexnet, can also using ZFNet or
GoogLeNet etc., herein without limitation, wherein by the last layer feature of network, (i.e. top-level feature, it reflects image
Further feature) it is evaluated for composition.
With obtained pattern data collection training deep learning network, study is allowed to the mapping between image and label.When
(for example, convergence criterion are as follows: pattern data concentrates the evaluation of each composition frame to export and the structure when deep learning network convergence
The label of picture frame is consistent;Or pattern data concentrates evaluation output and the ratio of the consistent composition frame of label to reach preset number), it obtains
To composition evaluation model.
Preferably, the further feature of network and at least one layer of shallow-layer feature are combined into use in deep learning network structure
It is evaluated in composition.Further feature, that is, top-level feature of network, shallow-layer are characterized in for opposite further feature, can be any non-top
Layer feature, such as first layer feature or second layer feature etc..Shallow-layer feature includes edge, color, brightness of image etc., deep layer
Characterizations integrality of picture material, such as the concordance of picture etc., are fused together shallow-layer feature and further feature
Can more complete picture engraving aesthetic feeling, to keep the evaluation of the composition evaluation model obtained based on deep learning method more acurrate.
Step S200 extracts the characteristic information of image in the frame of composition frame, and the composition frame includes photographic subjects.
The characteristic information includes that composition of geometry characterization factor, and/or the vision equilibrium characteristic factor, and/or human body are crucial
The integrity feature factor, and/or the expressive features factor, and/or background color terseness characterization factor of point.According to different type
Photography, such as personage shine, can choose composition of geometry characterization factor, the vision equilibrium characteristic factor, human body key point it is complete
Property characterization factor, the expressive features factor and background color terseness characterization factor;Pure natural land shines, and can choose geometry structure
Figure characterization factor.
Nine grids composition method is one of common composition rule of taking pictures, and is meant that, by composition frame with two vertical lines, two
After horizontal line is respectively divided into three parts, personage is placed on the crosspoint of four cut-off rules, this four points are exactly picture emphasis.
This composition method can make the picture entirety sense of equilibrium preferable, can also emphasize out to shoot the relativity between main body and secondary main body.
Based on the rule, according to the range information computational geometry structure in four crosspoints of nine grids in the center of photographic subjects and composition frame
Figure characterization factor, to measure whether composition meets nine grids composition rule.
Composition of geometry characterization factor can be further calculated using the following equation:
Wherein, f1 is composition of geometry characterization factor, (CRx,CRy) be photographic subjects centre coordinate, (Pix,Piy) it is target
The coordinate in four crosspoints of nine grids in composition frame, X and Y are the width and height of target pattern frame.
The vision equilibrium characteristic factor is mainly used for measuring the visual balance state of composition itself.When photographic subjects are for one
When, visual balance is mainly judged according to the facial orientation of photographic subjects;As shown in figure 9, facial orientation is to the right, the body of people exists
Position inside composition frame is left side, so being visually balance;If the body of people is on right side, consistent with facial orientation,
It is visually then unbalanced.
When photographic subjects are more people, in order to measure the reasonability of more people's space layouts, by human body detecting method, calculate
Everyone position of centre of gravity in composition;The position of centre of gravity of more people is obtained according to everyone position of centre of gravity, according to more people's
Location information of the position of centre of gravity in composition frame is to calculate the visual balance state of composition.If the position of centre of gravity of more people is in structure
The marginal position of picture frame is then non-equilibrium state;It is flat if the position of centre of gravity of more people is located at the central area of composition frame
Weighing apparatus state.For example, 3 people are on the left side of composition frame, for 1 people on the right of composition frame, personage's center of gravity is visually uneven on the left side
Weighing apparatus, this composition are bad.
The integrity feature factor of human body key point is primarily to butt occur, cutting hand, cut the feelings such as foot in punishment composition
Condition.When photographic subjects are personage, according to integrity degree of the human body key node of photographic subjects in composition frame, human body pass is obtained
The integrity feature factor of key point.Human body key point is extracted by image recognition, then judges human body key node in composition frame
Integrated degree.
The expressive features factor is used to react the expression of photographic subjects.When photographic subjects are personage, pass through recognition of face skill
Art determines that the expression of photographic subjects in composition frame, such as expression define three kinds of states: laughing at, is neutral, bad expression, to obtain table
Feelings characterization factor.
Background color terseness characterization factor is mainly used for measuring the succinct degree of background color.When photographic subjects are personage
When, according to the statistics with histogram of the background pixel of composition frame, obtain background color terseness characterization factor.According to background pixel
Statistics with histogram when pixel class is bigger, then shows as background face as a result, count the pixel class greater than preset quantity
Color is complicated, not enough succinctly.
For example, the pixel value in tri- channels background area RGB of composition frame is quantified respectively to 16 grades, so sharing 4096
Kind combination of pixels.The statistics with histogram for calculating pixel distribution, can obtain background color terseness characterization factor according to following formula
F5:
Wherein, S=i | His (i) >=γ hmaxIndicating that quantity accounting is greater than the background pixel set of preset ratio, γ is
Predetermined coefficient, hmaxFor the maximum statistics component of histogram, His (i) is the statistic of pixel i in histogram, | | S | | indicate collection
Close the element number of S.Be in when γ takes 0.01, f5 (0,1.5%] when, then background color is succinct;Otherwise not simple for background color
It is clean.
Step S310 evaluates the composition frame based on the random forest disaggregated model of the characteristic information, obtains institute
State the first assessment of composition frame.
Specifically, random forest classification is also a kind of machine learning algorithm, it is that one kind includes multiple decision trees, is based on majority
The classifier of voting mechanism, each decision tree are voted according to some features of input, and this feature needs Manual definition, selection
Most classification results vote as last prediction result.For example, a random forest grader comprising M decision tree,
New data is put into this M decision tree, every decision tree has a classification results, M classification results are obtained, to point
Class result is counted, using the most classification results of poll as last prediction result.
Random forest sorting algorithm is applied in the composition evaluation of composition frame, in conjunction with the feature of Manual definition, by big
The sample training of amount obtains the random forest disaggregated model evaluated for composition.The construction method of the model is as follows: obtaining several
It is added to the image of composition frame and the label of composition frame, obtains pattern data collection;According to the feature of Manual definition to composition number
Corresponding characteristic information is extracted according to each composition frame of collection;The characteristic information of each composition frame of pattern data collection and label is defeated
Enter random forest disaggregated model to be trained, is allowed to study to the mapping between characteristic information and label;When the disaggregated model is received
(for example, convergence criterion are as follows: the evaluation output for each composition frame is consistent with the label of the composition frame) when holding back, trained
Good random forest disaggregated model.
By taking personage is shone as an example, composition of geometry characterization factor, the vision equilibrium characteristic factor, the integrality of human body key point are selected
The characteristic informations such as characterization factor, the expressive features factor and background color terseness characterization factor, according to the people of features described above information
Work definition carries out corresponding feature information extraction to a composition frame to be evaluated, extracted characteristic information is inputted preparatory
The random forest disaggregated model of building obtains the evaluation output of the model, and evaluation output is commented as the first time of composition frame
Valence.
Step S400 evaluates the composition frame based on the composition evaluation model of deep learning method, obtains the structure
The second assessment of picture frame.
Step S500 obtains the overall merit of the composition frame according to the first assessment and the second assessment.
The first assessment of the present embodiment is obtained according to the feature and random forest disaggregated model of Manual definition, random gloomy
Standing forest class model is obtained by big data training, compared between the characteristic information defined according to artificial experience and composition evaluation
Mapping relations, it can obtain more accurate mapping relations, to improve the accuracy of the first assessment, and then improve synthesis
The accuracy of evaluation.
In another embodiment of the present invention, as shown in figure 3, a kind of photography composition evaluation method, comprising: with aforementioned reality
It applies example to compare, something in common no longer repeats, the difference is that after step S310, comprising:
Step S320 judges whether the first assessment meets preset condition;
Step S330 obtains the structure when the first assessment is unsatisfactory for preset condition, according to the first assessment
The overall merit of picture frame;
Step S410 is evaluated when the first assessment meets preset condition using the composition based on deep learning method
Model evaluates the composition frame, obtains the second assessment of the composition frame;
Step S510 obtains the overall merit of the composition frame according to the first assessment and the second assessment.
Specifically, if the first assessment is unsatisfactory for preset condition, such as when being lower than some pre-determined threshold, illustrate based on biography
The evaluation for feature of uniting is not high, and overall merit is not certainly high, so the second assessment need not be carried out, can reduce system in this way
Operand reduces system loading.
In one embodiment of the invention, as shown in figure 4, a kind of automatic photography method, comprising:
Step S600 generation includes the composition frame of photographic subjects.
Specifically, as needed, photographic subjects can be adult or children.It correspondingly, can be according to based on Identification of Images
Computer vision technique, generate include photographic subjects composition frame.At this point, with the image of photographic subjects, root in composition frame
The images such as the corresponding photo of photographic subjects can be exported according to the composition frame.
Step S700 uses any one photography composition evaluation method in aforementioned to evaluate the composition frame, obtains comprehensive
Close evaluation.
Step S800 controls the output of the composition frame according to the overall merit.
Step S900 photographs when photographic unit the receives composition frame according to the composition frame.
The present embodiment only when the overall merit of composition frame is high, just can further export the composition in automatic photography
Frame executes photography.It when overall merit is low, needs to regenerate composition frame, adjusts composition.It can be evaluated and be kept away by composition in this way
Exempt to export low-quality photo, and then low quality photo is avoided to occupy memory source, increase loss etc..
In order to improve automatic photography efficiency, multiple composition frames can be generated simultaneously, and photography is executed to each composition frame respectively
Composition evaluation, obtains the overall merit of each composition frame.If there is the overall merit of multiple composition frames reaches minimum output requirement,
The therefrom highest composition frame output of selection overall merit.If reaching minimum output requirement without composition frame, need to give birth to again
At composition frame.
In one embodiment of the invention, as shown in figure 5, a kind of photography composition evaluating apparatus 100, comprising:
Characteristic extracting module 110, the characteristic information of image in the frame for extracting composition frame, the composition frame include to clap
Take the photograph target.
First evaluation module 120 obtains the composition for evaluating based on the characteristic information the composition frame
The first assessment of frame.
Specifically, as needed, photographic subjects can be adult or children.It can be known using known face recognition technology
It whether include photographic subjects in other composition frame.But it include that the composition frames of photographic subjects is not necessarily a good composition, in order to
It avoids generating low quality photo, output as far as possible meets the photo that human aesthetic requires, and the composition to composition frame is needed to comment
Valence.
Feature information extraction is carried out to image in the frame of composition frame first.Characteristic information is usually to be passed through according to traditional photography
It tests, is mostly extracted from the bottom visual information of composition, and can be by limited or simple mathematical knowledge expression, for example, shooting
When target is personage, position feature of the photographic subjects in composition frame is (for example, the center and each side of composition frame that pass through photographic subjects
Range information reflection), size characteristic of the photographic subjects in composition frame.Position based on photographic subjects in composition frame is special
Sign, it can be determined that photographic subjects are the edges in center or composition frame, if being in the edge of composition frame, structure
Scheme undesirable;If opposite center, composition meet the aesthetic habit of people in picture;Based on photographic subjects in composition
Size characteristic in frame, it can be determined that whether photographic subjects size is too small in composition frame, if size is too small, composition is paid no attention to
Think.
Features above information is one or more mathematical expressions of parts of traditional photography experience, and this mathematical expression needs people
Work is pre-defined, and machine extracts corresponding characteristic information value according to the mathematical expression predetermined from specific composition frame,
The composition of composition frame is evaluated further according to these characteristic information values, obtains the first assessment.
Second evaluation module 130 comments the composition frame for the composition evaluation model based on deep learning method
Valence obtains the second assessment of the composition frame.
Specifically, manual features extraction comparison is laborious, and the feature that can manually extract is also than relatively limited, mostly shallow-layer
Feature has some features (such as further feature) that can not also express by limited mathematical knowledge, for example image is presented
Artistic conception, concordance of picture etc..
Deep learning is a kind of machine learning algorithm, and benefit is not need manually to extract feature, is learnt by supervised
Feature can be automatically extracted, shallow-layer feature can not only be extracted, moreover it is possible to extract further feature, pass through complicated mathematical knowledge table
Up to out.Deep learning network is a kind of neural network with more hidden layers, is with classical convolutional neural networks Alexnet
Example has 1 input layer, 6 hidden layers, 1 output layer, and totally 8 layers, every layer has thousands of a neurons, and every layer of output is the one of data
Kind feature, and feature hierarchy is improved with the increase of processing level, for example, first layer is the edge feature of image, the second layer is
The color characteristic of image, it is the concordance feature of image that third layer, which is the 8th layer of brightness ... ... of image, network it is former
Layer reflects the local feature of image, is called shallow-layer feature, such as the features such as edge, color, brightness of image, network it is last
One layer, i.e. the top layer global feature that reflects image, are called further feature, such as the concordance feature of image.
When being trained to deep learning network, only further feature can be evaluated applied to composition, it can also be by deep layer
Feature and shallow-layer feature combined application are evaluated in composition.When deep learning network convergence, composition evaluation model is obtained.Use this
Composition evaluation model evaluates composition frame, obtains the second assessment of the composition frame.
Overall merit module 140, for obtaining the composition frame according to the first assessment and the second assessment
Overall merit.
Specifically, the composition based on deep learning evaluates (i.e. the second assessment) cannot replace the first assessment completely, i.e.,
Composition evaluation model is set to consider further feature and shallow-layer feature, also can be problematic, the reason is that: one, deep learning learns out
Feature lack interpretation, it is why good, it is why bad, can not be explicitly stated;Two, the spy that deep learning learns out
Sign depends greatly on data set, and the distribution of data set and abundant degree will limit its effect;Exactly because above-mentioned two
There is uncontrollable situation, i.e. depth in reason, the generalization ability that will lead to the composition evaluation model based on deep learning method
Algorithm is practised to lack in common sense.
The first assessment is evaluated from the feature of Manual definition.The feature of Manual definition, i.e. traditional characteristic, mostly shallowly
Layer feature, interpretation be it is guaranteed, generalization ability is also more secure than deep learning, all for clearly good feature
Bad feature is specified, it is more suitable for going description with traditional characteristic, and can also largely reduce calculation amount.But it passes
System feature also has the problem of itself, is manually also difficult to define further feature, such as the concordance of whole picture etc. at present, and this
A little high-level things, exactly deep learning neural network are relatively good at.
The second assessment considers the further feature that machine deep learning arrives, and the two, which combines, just can more fully be commented
Valence had both considered interpretation feature, it is also considered that not describable further feature (i.e. high-level semantics features), it in this way could be complete
The quality for removing evaluation composition in face, existing details also have entirety, and generalization ability is more secure.
Overall merit is obtained according to the first assessment and the second assessment.It, just can be further only when overall merit is high
Composition frame is exported, is photographed;The promotion of patterning quality is promoted by the promotion that composition is evaluated, to avoid output low quality
Photo, and then avoid low quality photo occupy memory source, increase loss etc..
In one embodiment of the invention, as shown in fig. 6, a kind of photography composition evaluating apparatus 100, comprising:
Model construction module 150, for obtaining the label of several images for being added to composition frame and each composition frame,
Obtain pattern data collection;And with the pattern data collection training deep learning network, the deep layer of the deep learning network is special
It takes over for use and is evaluated in composition;When the deep learning network convergence, composition evaluation model is obtained.
Specifically, using large-scale dataset, such as millions of images, composition frame is added to every image, and stamp
Label obtains pattern data collection.Since composition evaluation is subjective, there are apparent individual differences, therefore, for composition frame
The mark of interior image is generally labeled same piece image using more people, takes proprietary average mark as image later
Final label.
Deep learning network structure using classical convolutional neural networks, such as Alexnet, can also using ZFNet or
GoogLeNet etc., herein without limitation, wherein by the last layer feature of network, (i.e. top-level feature, it reflects image
Further feature) it is evaluated for composition.
With obtained pattern data collection training deep learning network, study is allowed to the mapping between image and label.When
(for example, convergence criterion are as follows: pattern data concentrates the evaluation of each composition frame to export and the structure when deep learning network convergence
The label of picture frame is consistent;Or pattern data concentrates evaluation output and the ratio of the consistent composition frame of label to reach preset number), it obtains
To composition evaluation model.
Preferably, the further feature of network and at least one layer of shallow-layer feature are combined into use in deep learning network structure
It is evaluated in composition.Further feature, that is, top-level feature of network, shallow-layer are characterized in for opposite further feature, can be any non-top
Layer feature, such as first layer feature or second layer feature etc..Shallow-layer feature includes edge, color, brightness of image etc., deep layer
Characterizations integrality of picture material, such as the concordance of picture etc., are fused together shallow-layer feature and further feature
Can more complete picture engraving aesthetic feeling, to keep the evaluation of the composition evaluation model obtained based on deep learning method more acurrate.
Characteristic extracting module 110, the characteristic information of image in the frame for extracting composition frame, the composition frame include to clap
Take the photograph target.
The characteristic information includes that composition of geometry characterization factor, and/or the vision equilibrium characteristic factor, and/or human body are crucial
The integrity feature factor, and/or the expressive features factor, and/or background color terseness characterization factor of point.According to different type
Photography, such as personage shine, can choose composition of geometry characterization factor, the vision equilibrium characteristic factor, human body key point it is complete
Property characterization factor, the expressive features factor and background color terseness characterization factor;Pure natural land shines, and can choose geometry structure
Figure characterization factor.
Nine grids composition method is one of common composition rule of taking pictures, and is meant that, by composition frame with two vertical lines, two
After horizontal line is respectively divided into three parts, personage is placed on the crosspoint of four cut-off rules, this four points are exactly picture emphasis.
This composition method can make the picture entirety sense of equilibrium preferable, can also emphasize out to shoot the relativity between main body and secondary main body.
Based on the rule, according to the range information computational geometry structure in four crosspoints of nine grids in the center of photographic subjects and composition frame
Figure characterization factor, to measure whether composition meets nine grids composition rule.
Composition of geometry characterization factor can be further calculated using the following equation:
Wherein, f1 is composition of geometry characterization factor, (CRx,CRy) be photographic subjects centre coordinate, (Pix,Piy) it is target
The coordinate in four crosspoints of nine grids in composition frame, X and Y are the width and height of target pattern frame.
The vision equilibrium characteristic factor is mainly used for measuring the visual balance state of composition itself.When photographic subjects are for one
When, visual balance is mainly judged according to the facial orientation of photographic subjects;As shown in figure 9, facial orientation is to the right, the body of people exists
Position inside composition frame is left side, so being visually balance;If the body of people is on right side, consistent with facial orientation,
It is visually then unbalanced.
When photographic subjects are more people, in order to measure the reasonability of more people's space layouts, by human body detecting method, calculate
Everyone position of centre of gravity in composition;The position of centre of gravity of more people is obtained according to everyone position of centre of gravity, according to more people's
Location information of the position of centre of gravity in composition frame is to calculate the visual balance state of composition.If the position of centre of gravity of more people is in structure
The marginal position of picture frame is then non-equilibrium state;It is flat if the position of centre of gravity of more people is located at the central area of composition frame
Weighing apparatus state.For example, 3 people are on the left side of composition frame, for 1 people on the right of composition frame, personage's center of gravity is visually uneven on the left side
Weighing apparatus, this composition are bad.
The integrity feature factor of human body key point is primarily to butt occur, cutting hand, cut the feelings such as foot in punishment composition
Condition.When photographic subjects are personage, according to integrity degree of the human body key node of photographic subjects in composition frame, human body pass is obtained
The integrity feature factor of key point.Human body key point is extracted by image recognition, then judges human body key node in composition frame
Integrated degree.
The expressive features factor is used to react the expression of photographic subjects.When photographic subjects are personage, pass through recognition of face skill
Art determines that the expression of photographic subjects in composition frame, such as expression define three kinds of states: laughing at, is neutral, bad expression, to obtain table
Feelings characterization factor.
Background color terseness characterization factor is mainly used for measuring the succinct degree of background color.When photographic subjects are personage
When, according to the statistics with histogram of the background pixel of composition frame, obtain background color terseness characterization factor.According to background pixel
Statistics with histogram when pixel class is bigger, then shows as background face as a result, count the pixel class greater than preset quantity
Color is complicated, not enough succinctly.
For example, the pixel value in tri- channels background area RGB of composition frame is quantified respectively to 16 grades, so sharing 4096
Kind combination of pixels.The statistics with histogram for calculating pixel distribution, can obtain background color terseness characterization factor according to following formula
F5:
Wherein, S=i | His (i) >=γ hmaxIndicating that quantity accounting is greater than the background pixel set of preset ratio, γ is
Predetermined coefficient, hmaxFor the maximum statistics component of histogram, His (i) is the statistic of pixel i in histogram, | | S | | indicate collection
Close the element number of S.Be in when γ takes 0.01, f5 (0,1.5%] when, then background color is succinct;Otherwise not simple for background color
It is clean.
First evaluation module 120, for the random forest disaggregated model based on the characteristic information to the composition frame into
Row evaluation, obtains the first assessment of the composition frame.
Specifically, random forest classification is also a kind of machine learning algorithm, it is that one kind includes multiple decision trees, is based on majority
The classifier of voting mechanism, each decision tree are voted according to some features of input, and this feature needs Manual definition, selection
Most classification results vote as last prediction result.For example, a random forest grader comprising M decision tree,
New data is put into this M decision tree, every decision tree has a classification results, M classification results are obtained, to point
Class result is counted, using the most classification results of poll as last prediction result.
Random forest sorting algorithm is applied in the composition evaluation of composition frame, in conjunction with the feature of Manual definition, by big
The sample training of amount obtains the random forest disaggregated model evaluated for composition.The construction method of the model is as follows: obtaining several
It is added to the image of composition frame and the label of composition frame, obtains pattern data collection;According to the feature of Manual definition to composition number
Corresponding characteristic information is extracted according to each composition frame of collection;The characteristic information of each composition frame of pattern data collection and label is defeated
Enter random forest disaggregated model to be trained, is allowed to study to the mapping between characteristic information and label;When the disaggregated model is received
(for example, convergence criterion are as follows: the evaluation output for each composition frame is consistent with the label of the composition frame) when holding back, trained
Good random forest disaggregated model.
By taking personage is shone as an example, composition of geometry characterization factor, the vision equilibrium characteristic factor, the integrality of human body key point are selected
The characteristic informations such as characterization factor, the expressive features factor and background color terseness characterization factor, according to the people of features described above information
Work definition carries out corresponding feature information extraction to a composition frame to be evaluated, extracted characteristic information is inputted preparatory
The random forest disaggregated model of building obtains the evaluation output of the model, and evaluation output is commented as the first time of composition frame
Valence.
Second evaluation module 130 comments the composition frame for the composition evaluation model based on deep learning method
Valence obtains the second assessment of the composition frame.
Overall merit module 140, for obtaining the composition frame according to the first assessment and the second assessment
Overall merit.
The first assessment of the present embodiment is obtained according to the feature and random forest disaggregated model of Manual definition, random gloomy
Standing forest class model is obtained by big data training, compared between the characteristic information defined according to artificial experience and composition evaluation
Mapping relations, it can obtain more accurate mapping relations, to improve the accuracy of the first assessment, and then improve synthesis
The accuracy of evaluation.
In another embodiment of the present invention, as shown in fig. 7, a kind of photography composition evaluating apparatus 100, comprising: with it is preceding
Embodiment to be stated to compare, something in common no longer repeats, the difference is that, comprising:
Judgment module, judges whether the first assessment meets preset condition;
Overall merit module, for when the first assessment is unsatisfactory for preset condition, according to the first assessment
Obtain the overall merit of the composition frame;
Second evaluation module, for when the first assessment meets preset condition, using being based on deep learning method
Composition evaluation model the composition frame is evaluated, obtain the second assessment of the composition frame;
The overall merit module is further used for obtaining institute according to the first assessment and the second assessment
State the overall merit of composition frame.
Specifically, if the first assessment is unsatisfactory for preset condition, such as when being lower than some pre-determined threshold, illustrate based on biography
The evaluation for feature of uniting is not high, and overall merit is not certainly high, so the second assessment need not be carried out, can reduce system in this way
Operand reduces system loading.
In one embodiment of the invention, as shown in figure 8, a kind of electronic equipment 400, including memory 410 and processing
Device 420.The memory 410 is for storing computer program 430.The processor is realized such as when running the computer program
The photography composition evaluation method of foregoing description.
As an example, it is realized according to the step S200 of foregoing description extremely when processor 420 executes computer program
S500.Additionally, it is realized when processor 420 executes computer program each in the photography composition evaluating apparatus 100 of foregoing description
Module, the function of unit.As another example, processor 420 realized when executing computer program characteristic extracting module 110,
The function of first evaluation module 120, the second evaluation module 130 and overall merit module 140.
Optionally, according to specific needs of the invention are completed, the computer program can be divided into one or more
Module/unit.Each module/unit can be the series of computation machine program instruction section that can complete specific function.The calculating
Machine program instruction section is for describing implementation procedure of the computer program in photography composition evaluating apparatus 100.As an example,
The computer program can be divided into modules/unit in virtual bench, for example characteristic extracting module, the first evaluation
Module, the second evaluation module and overall merit module.
The processor is used for by executing the computer program to realize that photography composition is evaluated.As needed, institute
Stating processor can be central processing unit (CPU), graphics processing unit (GPU), digital signal processor (DSP), dedicated collection
At circuit (ASIC), field programmable gate array (FPGA), general processor or other logical devices etc..
The memory can be the internal storage unit and/or external storage that arbitrarily can be realized data, program storage
Equipment.For example, the memory can be plug-in type hard disk, intelligent memory card (SMC), secure digital (SD) card or flash card
Deng.The memory is used to store other programs and data of computer program, photography composition evaluating apparatus 100.
The electronic equipment 400 can be any computer equipment, for example desktop PC (desktop), portable
Computer (laptop), palm PC (PDA) or server (server) etc..As needed, the electronic equipment 400 can be with
Including input-output equipment, display equipment, network access equipment and bus 440 etc..The electronic equipment 400 can also be monolithic
Machine, or it is integrated with the calculating equipment of central processing unit (CPU) and graphics processing unit (GPU).
It for realizing the unit of corresponding function, the division of module is it will be appreciated by persons skilled in the art that above-mentioned
Said units, module are further divided or combined, i.e., according to application demand in convenient explanation, the purpose of narration
Device/equipment internal structure is re-started into division, combination, with the function of the above-mentioned record of realization.In above-described embodiment
Individual physical unit can be respectively adopted in each unit, module, two or more units, module can also be integrated
In a physical unit.Each unit, module in above-described embodiment can be using the realizations of hardware and/or SFU software functional unit
Corresponding function.Between multiple units, component, module in above-described embodiment can with direct-coupling, INDIRECT COUPLING or communication
Connection can be realized by bus or interface;Coupling, connection between multiple units or device can be electrical, mechanical
Or similar mode.Correspondingly, each unit in above-described embodiment, the specific name of module also only to facilitate narration and
It distinguishes, and does not have to the protection scope of limitation the application.
In one embodiment of the invention, a kind of computer readable storage medium is stored thereon with computer program, institute
The photography composition evaluation method recorded such as previous embodiment can be realized by stating when computer program is executed by processor.It that is to say, when
Some or all of technical solution that the aforementioned embodiment of the present invention contributes to the prior art passes through computer software product
When mode emerges from, aforementioned computer software product is stored in a computer readable storage medium.The computer can
Reading storage medium can be any portable computers program code entity apparatus or equipment.For example, described computer-readable to deposit
Storage media can be USB flash disk, mobile disk, magnetic disk, CD, computer storage, read-only memory, random access memory etc..
In one embodiment of the invention, as shown in Figure 10, Figure 11, a kind of automatic photographing device 500, comprising:
Composition frame generation module 200, for generate include photographic subjects composition frame.
Photography composition evaluating apparatus 100 described in aforementioned any embodiment obtains described for evaluating composition frame
The overall merit of composition frame.
Composition frame output module 300, for controlling the output of the composition frame according to the overall merit.
Photographic unit 502, for being photographed according to the composition frame when receiving the composition frame.Described the ministry of photography
Part 502 is used to optical imagery shooting being fixed on corresponding storage medium.Optionally, the photographic unit 502 can be number
Code camera.
In order to promote the application performance of automatic photographing device 500, automatic photographing device 500 further includes pedestal 504 and optics
Camera lens 252.Pedestal 504 is used to support the other structures of automatic shooting equipment 500, component etc..Pedestal 504 is generally cylindrical,
Its circumferential direction is circumferencial direction.Optical lens 252 is set to the periphery wall of pedestal 504.Optical lens 252 is used for will be to be captured
Target carry out optical imagery.The optical lens 252 is wide-angle lens, has the visual field of the camera lens 503 greater than photographic unit 502
Angle.I.e. the field angle of photographic unit 502 is determined by the field angle of its included camera lens 503.The camera lens 503 of photographic unit 502, also
It is a kind of optical lens, is arranged on bracket component 506.In order to capture photographic subjects in spatial dimension as big as possible, three
Optical lens 252 is intervally arranged along the circumferencial direction of pedestal 504.The field angle of three optical lens 252 is adjacent to each other or again
It is folded, so as to obtain the photographic subjects in 360 ° of spaces of level.
Automatic photographing device 500 further includes imaging sensor 254.The quantity of imaging sensor 254 can be with optical lens 252
It is identical, and the assembling cooperation one by one that corresponds to each other.Optical picture of the imaging sensor 254 for that will enter in corresponding optical lens 252
As signal is converted to the picture signal of electronic data, and the picture signal of the electronic data is exported in order to subsequent place
Reason.In Figure 11, imaging sensor 254 is arranged in pedestal 504, and explanation dotted line illustrates its substantially installation position for the ease of illustration
It sets.
In order to facilitate that the activity to photographic unit 502 is being propped up to higher image quality, the setting of photographic unit 502 is obtained
On frame component 506, and it can be rotatably set relative to bracket component 506.Specifically, photographic unit 502 around horizontal extension and hangs down
Directly in 602 pitch rotation of the pivot center of shooting direction, the activity of vertical up and down direction can be realized.Photographic unit 502 is opposite
It can be rotatably set in the pivot center 603 of bracket component 506 extended along shooting direction.When photographic unit 502 rotates 90 °
When, " perpendicular to clap " may be implemented.Bracket component 506 can be rotatably set relative to the axis 601 that vertical direction extends, so as to
Photographic unit 502 is driven to realize the rotation within the scope of 360 ° of horizontal view angle.Wherein, appoint in three pivot centers 601,602,603
Meaning two perpendicular to one another, to be configured to solid space coordinate system.
Photography composition evaluating apparatus 100, composition frame generation module 200, composition frame output module 300 are arranged in pedestal 504
It is interior.Aforementioned electronic devices 400 also can be set in pedestal 504.
The image of higher shooting quality is obtained by aforementioned photography composition evaluation method, automatic shooting method, device
Substantially steps are as follows:
Photographic subjects are captured on a large scale by optical lens 252, after finding photographic subjects, obtain the position letter of photographic subjects
Breath adjusts photographic unit 502, it is made to be directed at photographic subjects, has one to include shooting mesh in the visual field of photographic unit 502 at this time
Target image, in order to further enhance the picture quality shot to photographic subjects, composition frame generation module 200 is to current figure
Photographic subjects setting composition frame (preview pane in similar digital camera) as in, obtains initial pictures (the i.e. structure of photographic subjects
Image in picture frame).There are many setting means of composition frame, such as centered on photographic subjects, takes the external envelope of photographic subjects,
External envelope is arranged by the photographic scale that camera is shot again, obtains a kind of composition frame.Photography composition evaluating apparatus 100 is to this
Composition frame is patterned evaluation, obtains the overall merit of composition frame, that is, determines whether the initial pictures meet photography composition requirement;
When the overall merit meets photography composition requirement, triggering composition frame output module 300 exports the composition frame, photographic unit 502
When receiving the composition frame, then execute movement of taking pictures.When the overall merit is unsatisfactory for photography composition requirement, repetition pair can be passed through
The adjusting of photographic unit 502, the generation and evaluation of composition frame until output meets the composition frame of photography composition requirement, then execute
It takes pictures.It takes pictures due to only just will do it when initial pictures meet photography composition requirement, so the figure of better quality can be obtained
Picture.
It can one of in the following manner or a variety of adjusting photographic units 502: the adjusting of photographic unit enlargement ratio
(zoom), photographic unit with respect to the horizontal plane on pitch regulation and photographic unit in moving in parallel on the water surface and/or phase
The spatial position that axis around vertical direction horizontally rotates is adjusted, that is to say and carry out PTZ (Pan/Tilt/Zoom) operation.
In order to improve automatic photography efficiency, multiple composition frames can be generated simultaneously, and photography is executed to each composition frame respectively
Composition evaluation, obtains the overall merit of each composition frame.If there is the overall merit of multiple composition frames reaches minimum output requirement,
The therefrom highest composition frame output of selection overall merit.If reaching minimum output requirement without composition frame, needs to adjust and take the photograph
Shadow component regenerates composition frame.
It should be noted that above-described embodiment can be freely combined as needed.The above is only of the invention preferred
Embodiment, it is noted that for those skilled in the art, in the premise for not departing from the principle of the invention
Under, several improvements and modifications can also be made, these modifications and embellishments should also be considered as the scope of protection of the present invention.
Claims (16)
1. a kind of photography composition evaluation method characterized by comprising
The characteristic information of image in the frame of composition frame is extracted, the composition frame includes photographic subjects;
The composition frame is evaluated based on the characteristic information, obtains the first assessment of the composition frame;
The composition frame is evaluated based on the composition evaluation model of deep learning method, obtains second of the composition frame
Evaluation;
According to the first assessment and the second assessment, the overall merit of the composition frame is obtained.
2. photography composition evaluation method according to claim 1, which is characterized in that the structure based on deep learning method
Figure evaluation model includes: before being evaluated the composition frame
The label for obtaining several images for being added to composition frame and each composition frame, obtains pattern data collection;
It is evaluated with the further feature of the pattern data collection training deep learning network, the deep learning network for composition;
When the deep learning network convergence, composition evaluation model is obtained.
3. photography composition evaluation method according to claim 2, which is characterized in that the deep learning network
Further feature is evaluated for composition further include:
The further feature of the deep learning network and at least one shallow-layer feature are combined to be evaluated for composition.
4. photography composition evaluation method according to claim 1, which is characterized in that described to be based on the characteristic information to institute
It states composition frame and evaluate and include:
The composition frame is evaluated based on the random forest disaggregated model of the characteristic information.
5. photography composition evaluation method according to claim 1, which is characterized in that image in the frame for extracting composition frame
Characteristic information include:
According to the range information in four crosspoints of nine grids in the center of photographic subjects and composition frame, composition of geometry feature is obtained
The factor;And/or
When photographic subjects is one, the vision equilibrium characteristic factor is obtained according to the facial orientation of the photographic subjects;And when
When photographic subjects are more people, the vision equilibrium characteristic factor is obtained according to personage's center of gravity of the photographic subjects;And/or
When photographic subjects are personage, according to integrity degree of the human body key node of the photographic subjects in composition frame, obtain
The integrity feature factor of human body key point;And/or
When photographic subjects are personage, according to the expression of the photographic subjects, the expressive features factor is obtained;And/or
When photographic subjects are personage, according to the statistics with histogram of the background pixel of composition frame, it is special to obtain background color terseness
Levy the factor.
6. a kind of photography composition evaluation method characterized by comprising
The characteristic information of image in the frame of composition frame is extracted, the composition frame includes photographic subjects;
The composition frame is evaluated based on the characteristic information, obtains the first assessment of the composition frame;
Judge whether the first assessment meets preset condition;
When the first assessment is unsatisfactory for preset condition, commented according to the synthesis that the first assessment obtains the composition frame
Valence;
When the first assessment meets preset condition, using the composition evaluation model based on deep learning method to the structure
Picture frame is evaluated, and the second assessment of the composition frame is obtained;And it described comments according to the first assessment and for the second time
Valence obtains the overall merit of the composition frame.
7. a kind of automatic photography method characterized by comprising
Generation includes the composition frame of photographic subjects;
Photography composition evaluation method according to claim 1-6 evaluates the composition frame, is integrated
Evaluation;
The output of the composition frame is controlled according to the overall merit;
When photographic unit receives the composition frame, photograph according to the composition frame.
8. a kind of photography composition evaluating apparatus characterized by comprising
Characteristic extracting module, the characteristic information of image in the frame for extracting composition frame, the composition frame includes photographic subjects;
First evaluation module obtains the of the composition frame for evaluating the composition frame based on the characteristic information
Primary evaluation;
Second evaluation module is evaluated the composition frame for the composition evaluation model based on deep learning method, is obtained
The second assessment of the composition frame;
Overall merit module, for obtaining the synthesis of the composition frame according to the first assessment and the second assessment
Evaluation.
9. photography composition evaluating apparatus according to claim 8, which is characterized in that further include:
Model construction module obtains composition for obtaining the label of several images for being added to composition frame and each composition frame
Data set;And with the pattern data collection training deep learning network, the further feature of the deep learning network is used for structure
Figure evaluation;When the deep learning network convergence, composition evaluation model is obtained.
10. photography composition evaluating apparatus according to claim 9, it is characterised in that:
The model construction module is further used for the pattern data collection training deep learning network, the deep learning
The further feature of network and at least one shallow-layer feature are combined to be evaluated for composition.
11. photography composition evaluating apparatus according to claim 8, it is characterised in that:
First evaluation module is further used for the random forest disaggregated model based on the characteristic information to the composition frame
It is evaluated.
12. photography composition evaluating apparatus according to claim 8, it is characterised in that:
The characteristic extracting module is further used for four crosspoints according to nine grids in the center of photographic subjects and composition frame
Range information, obtain composition of geometry characterization factor;And/or
When photographic subjects is one, the vision equilibrium characteristic factor is obtained according to the facial orientation of the photographic subjects;And when
When photographic subjects are more people, the vision equilibrium characteristic factor is obtained according to personage's center of gravity of the photographic subjects;And/or
When photographic subjects are personage, according to integrity degree of the human body key node of the photographic subjects in composition frame, obtain
The integrity feature factor of human body key point;And/or
When photographic subjects are personage, according to the expression of the photographic subjects, the expressive features factor is obtained;And/or
When photographic subjects are personage, according to the statistics with histogram of the background pixel of composition frame, it is special to obtain background color terseness
Levy the factor.
13. a kind of photography composition evaluating apparatus characterized by comprising
Characteristic extracting module, the characteristic information of image in the frame for extracting composition frame, the composition frame includes photographic subjects;
First evaluation module obtains the of the composition frame for evaluating the composition frame based on the characteristic information
Primary evaluation;
Judgment module, judges whether the first assessment meets preset condition;
Overall merit module, for being obtained according to the first assessment when the first assessment is unsatisfactory for preset condition
The overall merit of the composition frame;
Second evaluation module, for when the first assessment meets preset condition, using the structure based on deep learning method
Figure evaluation model evaluates the composition frame, obtains the second assessment of the composition frame;
The overall merit module is further used for obtaining the structure according to the first assessment and the second assessment
The overall merit of picture frame.
14. a kind of automatic photographing device characterized by comprising
Composition frame generation module, for generate include photographic subjects composition frame;
The described in any item photography composition evaluating apparatus of claim 8-13 obtain comprehensive for evaluating the composition frame
Close evaluation;
Composition frame output module, for controlling the output of the composition frame according to the overall merit;
Photographic unit, for being photographed according to the composition frame when receiving the composition frame.
15. a kind of electronic equipment characterized by comprising
Memory, for storing computer program;
Processor realizes photography composition according to any one of claim 1 to 6 when for running the computer program
Evaluation method.
16. a kind of computer readable storage medium, is stored thereon with computer program, it is characterised in that:
The computer program realizes that photography composition according to any one of claim 1 to 6 is commented when being executed by processor
Valence method.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910708006.7A CN110519509A (en) | 2019-08-01 | 2019-08-01 | Composition evaluation method, method for imaging, device, electronic equipment, storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910708006.7A CN110519509A (en) | 2019-08-01 | 2019-08-01 | Composition evaluation method, method for imaging, device, electronic equipment, storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110519509A true CN110519509A (en) | 2019-11-29 |
Family
ID=68624210
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910708006.7A Pending CN110519509A (en) | 2019-08-01 | 2019-08-01 | Composition evaluation method, method for imaging, device, electronic equipment, storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110519509A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111507941A (en) * | 2020-03-24 | 2020-08-07 | 杭州电子科技大学 | Composition characterization learning method for aesthetic quality evaluation |
WO2022262763A1 (en) * | 2021-06-18 | 2022-12-22 | 影石创新科技股份有限公司 | Image composition quality evaluation method and apparatus |
CN118334611A (en) * | 2024-06-13 | 2024-07-12 | 成都航空职业技术学院 | Multi-mode image data real-time fusion method based on deep learning |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101415077A (en) * | 2007-10-17 | 2009-04-22 | 索尼株式会社 | Composition determining apparatus, composition determining method, and program |
CN104346801A (en) * | 2013-08-02 | 2015-02-11 | 佳能株式会社 | Image-composition evaluating device, information processing device and method thereof |
CN106097358A (en) * | 2016-06-17 | 2016-11-09 | 广州市幸福网络技术有限公司 | Image background complexity detection method and system |
CN107087427A (en) * | 2016-11-30 | 2017-08-22 | 深圳市大疆创新科技有限公司 | Control method, device and the equipment and aircraft of aircraft |
CN107610123A (en) * | 2017-10-11 | 2018-01-19 | 中共中央办公厅电子科技学院 | A kind of image aesthetic quality evaluation method based on depth convolutional neural networks |
CN108366203A (en) * | 2018-03-01 | 2018-08-03 | 北京金山安全软件有限公司 | Composition method, composition device, electronic equipment and storage medium |
US20180367752A1 (en) * | 2015-12-30 | 2018-12-20 | Google Llc | Low Power Framework for Controlling Image Sensor Mode in a Mobile Image Capture Device |
CN109657539A (en) * | 2018-11-05 | 2019-04-19 | 深圳前海达闼云端智能科技有限公司 | Face value evaluation method and device, readable storage medium and electronic equipment |
CN109697399A (en) * | 2017-10-24 | 2019-04-30 | 普天信息技术有限公司 | A kind of facial expression recognizing method and device |
-
2019
- 2019-08-01 CN CN201910708006.7A patent/CN110519509A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101415077A (en) * | 2007-10-17 | 2009-04-22 | 索尼株式会社 | Composition determining apparatus, composition determining method, and program |
CN104346801A (en) * | 2013-08-02 | 2015-02-11 | 佳能株式会社 | Image-composition evaluating device, information processing device and method thereof |
US20180367752A1 (en) * | 2015-12-30 | 2018-12-20 | Google Llc | Low Power Framework for Controlling Image Sensor Mode in a Mobile Image Capture Device |
CN106097358A (en) * | 2016-06-17 | 2016-11-09 | 广州市幸福网络技术有限公司 | Image background complexity detection method and system |
CN107087427A (en) * | 2016-11-30 | 2017-08-22 | 深圳市大疆创新科技有限公司 | Control method, device and the equipment and aircraft of aircraft |
CN107610123A (en) * | 2017-10-11 | 2018-01-19 | 中共中央办公厅电子科技学院 | A kind of image aesthetic quality evaluation method based on depth convolutional neural networks |
CN109697399A (en) * | 2017-10-24 | 2019-04-30 | 普天信息技术有限公司 | A kind of facial expression recognizing method and device |
CN108366203A (en) * | 2018-03-01 | 2018-08-03 | 北京金山安全软件有限公司 | Composition method, composition device, electronic equipment and storage medium |
CN109657539A (en) * | 2018-11-05 | 2019-04-19 | 深圳前海达闼云端智能科技有限公司 | Face value evaluation method and device, readable storage medium and electronic equipment |
Non-Patent Citations (1)
Title |
---|
温坤哲等: ""深度卷积神经网络在图像美学评价的应用综述"", 《计算机工程与应用》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111507941A (en) * | 2020-03-24 | 2020-08-07 | 杭州电子科技大学 | Composition characterization learning method for aesthetic quality evaluation |
WO2022262763A1 (en) * | 2021-06-18 | 2022-12-22 | 影石创新科技股份有限公司 | Image composition quality evaluation method and apparatus |
CN118334611A (en) * | 2024-06-13 | 2024-07-12 | 成都航空职业技术学院 | Multi-mode image data real-time fusion method based on deep learning |
CN118334611B (en) * | 2024-06-13 | 2024-08-20 | 成都航空职业技术学院 | Image data real-time processing method based on deep learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7149692B2 (en) | Image processing device, image processing method | |
CN113065558A (en) | Lightweight small target detection method combined with attention mechanism | |
JP5818799B2 (en) | Estimating the aesthetic quality of digital images | |
WO2020134528A1 (en) | Target detection method and related product | |
US9582731B1 (en) | Detecting spherical images | |
WO2019210555A1 (en) | People counting method and device based on deep neural network and storage medium | |
CN110807759B (en) | Method and device for evaluating photo quality, electronic equipment and readable storage medium | |
CN112040834A (en) | Eyeball tracking method and system | |
CN108319953A (en) | Occlusion detection method and device, electronic equipment and the storage medium of target object | |
CN112733802B (en) | Image occlusion detection method and device, electronic equipment and storage medium | |
CN108093174A (en) | Patterning process, device and the photographing device of photographing device | |
US20110292051A1 (en) | Automatic Avatar Creation | |
US20180357819A1 (en) | Method for generating a set of annotated images | |
CN104751407B (en) | A kind of method and apparatus for being blurred to image | |
US9106838B2 (en) | Automatic photographing method and system thereof | |
CN110519509A (en) | Composition evaluation method, method for imaging, device, electronic equipment, storage medium | |
CN109816745A (en) | Human body thermodynamic chart methods of exhibiting and Related product | |
WO2020211427A1 (en) | Segmentation and recognition method, system, and storage medium based on scanning point cloud data | |
JP2006119817A (en) | Image processor | |
US20160117834A1 (en) | Automatic, computer-based detection of triangular compositions in digital photographic images | |
CN113435282A (en) | Unmanned aerial vehicle image ear recognition method based on deep learning | |
CN104751406A (en) | Method and device used for blurring image | |
US8170299B2 (en) | Image output method, image output device, and image output program | |
CN109587394A (en) | A kind of intelligence patterning process, electronic equipment and storage medium | |
US20200019764A1 (en) | Portrait Image Evaluation Based on Aesthetics |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20191129 |